diff --git a/.gitattributes b/.gitattributes index 6492d3d34f22fd39673883831704ef7624d2a79e..687381f0fdc24de0bd6a6ed0bfe5849e28bc310c 100644 --- a/.gitattributes +++ b/.gitattributes @@ -60,3 +60,93 @@ HVU_QA/t5-viet-qg-finetuned/checkpoint-9000/spiece.model filter=lfs diff=lfs mer HVU_QA/t5-viet-qg-finetuned/checkpoint-9500/model.safetensors filter=lfs diff=lfs merge=lfs -text HVU_QA/t5-viet-qg-finetuned/checkpoint-9500/optimizer.pt filter=lfs diff=lfs merge=lfs -text HVU_QA/t5-viet-qg-finetuned/checkpoint-9500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/spiece.model filter=lfs diff=lfs merge=lfs -text diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f65fd350aeef143acf0fd1340930d3a33dc62dc8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf586da4797137e1b3d2df95c2878d8e6fd68cc32c20b35b036a5e9d4e4ecee1 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..59f78cd81c980e18f6d20a895d99da9f06e6edba --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f6aa749d6a6560fc42920e4bdc94c5d786e2e9a742630a9e2009d44c0c0ee40 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..3ffbc513adf98c13f592f386b68c06faff8304ae Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..b9e0bb649647703c6838e23b9a1c7386fc3e744c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..cfba86ebfc652b9717a970ee039aa4116ff7d64b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/trainer_state.json @@ -0,0 +1,7034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.3861152940267964, + "eval_steps": 500, + "global_step": 10000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6089578905600000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dcfd39411ca2e03af25e59f7b4c13c401760fce7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73cd51e828fe49387b89507a70438ff9d003f424cfa44b90899a7a7b5ae6b15d +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..780f1e70fa60ec90ea219dfe21d5910b6f3b7689 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4693ce5546cfd8ab7a02ee4233bbee17d28e94f8e0ee770ded1c245b8aa44c8 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..e4ccbaef380e5412144694a0547f34ae94918ca0 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..0918ac8c1536d0159d4b0fe71740b34031957565 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..050bd19968ea8eaafa462a54af3168cdfedb0388 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/trainer_state.json @@ -0,0 +1,7384 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4054210587281362, + "eval_steps": 500, + "global_step": 10500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6394057850880000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ec249917c8bb642a6749cd13c32e0101046f36d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:feef16fdfd3bf1b51a40e37272e9a09256f7a9ebfe58d9277145febef4fb0c7f +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..0e405bba9e7a32aa89e9c7895c2d21bb2f947156 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47e431aa70fbf8d5270aa73de0a0df788c50e275398015231f20abf6e37e557a +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..152337bed29eafa2466026ea135e73d93694032c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..b816873c93993f3a8ff00fd3ef371b8606d24fc7 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..85537e1cc554085fd2385e320071246b60dc33da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/trainer_state.json @@ -0,0 +1,7734 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.42472682342947604, + "eval_steps": 500, + "global_step": 11000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6698536796160000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c49538257d01883764dda8736679405cf86ab3e2 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49ba6d6bf9a36d4b97949dfba701fb3e1cc67aa2e91b74ca3d2c8b50cb985022 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..6a4ac1df7557cafd050d175ff5763d069187312e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08adf606a2c74527c040d50484a1e0a069a5e68bfb86178fd1aee08ef5ba957e +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..a6c563865a10c6ed2f0a09d2de00c0c8982535f8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..815c16b37ae2aa56ed9b677ff5433d442f8b44d3 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b437c593cd841ba7da4422d1272f7203f0e9f257 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/trainer_state.json @@ -0,0 +1,8084 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.44403258813081586, + "eval_steps": 500, + "global_step": 11500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 7003015741440000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-11500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6bf7c4a0a853d16b5e171530f6048696b0fcb3fc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f3ca10fb4146e80c7105189e46459bc0a2ea2bef5e29789d356d9cf125194ec +size 290455552 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7c112cc9ca2b8ee620a148ae1cb38c5b0ad59d21 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1473ca5dd1863a1b1e11f13099992f2fd0513bf88bcdd548e456ae885e0a4bd +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..61af2c10df4888045274c72a6ff85c5bce4cd2f6 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..4be94e4cd9bc51c60e0228636f64d7f6b825a8ac Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..55d4b8be745cd889e0431e77f5c3f75c082b4e73 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/trainer_state.json @@ -0,0 +1,8434 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4633383528321557, + "eval_steps": 500, + "global_step": 12000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 7307494686720000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..021367e6ebfc1dde19f03765c0701e47439de5a3 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d699df2e5b536164397a46f4482120cf072d80b4cbc8fdf8dc291dcb645595ad +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..08a258eae3699d3f5a2f4eb0b71202635841956d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99e76f7de5f05c102c45e725faf797dfd0006c185c514412004509cad08b3a68 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..1cb5823528912cb8fd3313778c755408540b8ed2 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ee06cf4f6c7999ff2a02e6cf3c8a36b89c5773f1 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..843b367303638a4875df480109bb56dad333ba6f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/trainer_state.json @@ -0,0 +1,8784 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4826441175334955, + "eval_steps": 500, + "global_step": 12500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 7611973632000000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-12500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8b7415186e16348b16e7940b9b81f0e7d1bb437c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6499e2f12d34cd4c4a2d39e34b8843fa4d6e299d6c83e29aad62df6be95bb12 +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..021d820cae9992330d4ce4f85ad6d11563322674 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:163085afd0fe15cbd07618619fe2841b261bea9a1be9290a6f16328e562793ee +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..6d3043aaa0c762470dac4378f84075e85cdad1d7 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..cb9fd959d2043495385c2ceaed043ecd466eaf85 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..fa02f5f0dfa2504442177b2e0cbd44f0a632380a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/trainer_state.json @@ -0,0 +1,9134 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5019498822348353, + "eval_steps": 500, + "global_step": 13000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 7916452577280000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b928c6bd1c29e8c14f1b74edd8bbb235cde2e941 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecd957e147df4c1660d07c3d4021fd53ee7eeed4fd5ef11ce09fcbb2af5653d2 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4644ea075199a453fc9d18cb9b9bd8bbbe18e4fe --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c72dd71962e9295a52194d148b489e8f8efe8c9f18268e180a2a7efe0f31edc +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c91fe787571b7e5350caed11724029ad831c446e Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..3f3ffc67603271c356ff37e7e8bdf249db1370ac Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..7dd5719957db08d3bc1da64a259172b635d8fff3 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/trainer_state.json @@ -0,0 +1,9484 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5212556469361751, + "eval_steps": 500, + "global_step": 13500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 8220931522560000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-13500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0a0d89b3249ba38cd3e3e33ca6bb1b9e9e776624 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f23ab952d15e507df07b769106f174a1b044f4b39291403de245ce5f59bbfc1 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..2a381c85ed896701e8580ab6953b20a8fc60ccd3 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0852e33828dc35ab5cd8ddc85012adabeff4ac926a4f8d2d1ce597f04ee8867 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..0532c45e77ed311c801147ec1f1b236500347a12 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..319beaee46500ee8bc59f347b65b2f2585104ab4 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d927bb0112464224e5c1457829326b73cfc23f42 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/trainer_state.json @@ -0,0 +1,9834 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5405614116375149, + "eval_steps": 500, + "global_step": 14000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 8525410467840000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5613066248387591aefe2a7e714264e00092b778 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36cdc5529cb236d2ef1d53ce46c4d1ef9aaee7df881bf1a624a74cf6d05ce7c0 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..72c4e454bca15d151016e08fb21260e92deb420b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aab5621eb593f9ef615bffc8117a1572affb2efb57ecc986b2bbcb7c942fecbb +size 292290560 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..36029cfd817456049baadcbc7fe8396abe6f7463 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..8ff7e6d3eef7bdf5abbc13b84474d023d6cb82c0 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..f42376e8b29e5b9282134d601ea753616d1fcddb --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/trainer_state.json @@ -0,0 +1,10184 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5598671763388547, + "eval_steps": 500, + "global_step": 14500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 8829889413120000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-14500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ae9527e0612abf7cdef5a705a353b54cd0ec6805 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9b58aa0dbddbbaf2f3d7c3c5265afe20652e7a16cfe542ed2b234c404b4d497 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7cd81afdce0359d841cca646df53f584ee7193e5 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c0ac73ddf9101f197629ce5b5ddb8cd412715a03b83a144e72dd9f7e4069125 +size 290455552 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..cff21f3e5e345dccdda399c63e8d59668e8035c4 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6dba558e667db9be61d21e561232711b8c214555 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..8dd79404e113ed6c8bb3cf31e8e3c3d3b1297492 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/trainer_state.json @@ -0,0 +1,10534 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5791729410401946, + "eval_steps": 500, + "global_step": 15000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 9134368358400000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..20dc62f9375568ed78b1b4eded9ee6c02da5586b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f46643b72cd3a10deb9147d29588ac7d3398df49fbfcbfbea903a25db0d54050 +size 290455552 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5b9504ab01589eacf0731bc3f3dba2f98767bd94 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a46377bbe4a964c3e3da1ce5a4940dbe2ccfe3e57713aaa0775617939afd583 +size 292028416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..253e90c2d5120224e6d4049bec5fc542ab7f9bf2 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..2d81f98e1616b2f1116c22bd2727e818825d4eaf Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..ea0d62e5e4feb93f29d9eac74ea41834d0473378 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/trainer_state.json @@ -0,0 +1,10884 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5984787057415344, + "eval_steps": 500, + "global_step": 15500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 9438847303680000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-15500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4312732bfc9661ec36606f5469e21709e9eeef51 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf069e1a7d0c22ee935c8dc77abda318ffeb20aa8b09f8f0eb3ba267681acd74 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..165ea09780f32ea1be5c98591ccd35e52c99ec49 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e89b03bc9ae4974ce49314da3c5b0105b5720fd8a0b36e70c58a082f48011f0 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..42ff2f3723f7f12e3a129b585000eef178553ebc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..020a440aa49d94af48fd4e97c0d4ab0071f2c9e1 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..91a8ee13c40ade6411fddb9dc5fe416d691fdd87 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/trainer_state.json @@ -0,0 +1,11234 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6177844704428742, + "eval_steps": 500, + "global_step": 16000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 9743326248960000.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b1deba41925dd08a45d0fad738c7f4beaf3df39d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8494a30952f9997d74e1b080032554c7c7a764d232a4fc8d04e279d64fc6313b +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d31ea93bc1ed38567f220ce67911ee71610f6ea2 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6849a0ba4c06342598a9e854ff891cde40c3b9b6754af117c0eed569da48ba93 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..52ffb367d51442fdb29cae2f8e7c0ca82142f418 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f6c3c3d9027fa246c4bced3f58e594b129cf865b Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..821662b9ad2def68fe25563f777c026b752313ac --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/trainer_state.json @@ -0,0 +1,11584 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.637090235144214, + "eval_steps": 500, + "global_step": 16500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.004780519424e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-16500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f50c0ba5d4f974b893464e6700f4165d01a22e5a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:318abcf10167ce1668a9d7ca7d36e3a7963f050b0281b787740c953e39bc3319 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..1d7d6ba3918d4ce4a8ed8db484b0358df5a1dcb9 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97bc76d555eea9533334ef4fbf07d2ad900a6f76d0d589c0f301380f08303350 +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c524e692a3adb900bb8a4884c514b13efb77e3c8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..c47370faad894a32db651597b537bf1a8382469c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..6490a48f8d32b803fe96936cac440326a3c512c7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/trainer_state.json @@ -0,0 +1,11934 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6563959998455539, + "eval_steps": 500, + "global_step": 17000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.035228413952e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0c38759c9a313d68718b6095653d137b3ef6c4b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf702d2da3ed440115b4e300cebfeb859a2934f61a77fc8fbe30ffdebd6bda74 +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7284ac458a83f91f1f42dd4f8a883b361f26293f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4cbeb85c508429194ff39ee73fd59198dbed02cdf83d273966a9773de4c8c11 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c91f83a1481c9642b6b3e2b5abacecceb40f35c0 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1148bda1f49dd1210343994220810b5cab933e28 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..a24d0051d2407936770e1e2c5b93c0be7b6df054 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/trainer_state.json @@ -0,0 +1,12284 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6757017645468937, + "eval_steps": 500, + "global_step": 17500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.06567630848e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-17500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..64a1d60d51d5984a8fa939f67df6cd71ea6a7fd5 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ad8098999213cec1db2d0c079f96680609e471266cdadb295f0a40a12d53d4e +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..ed51dd35ace5342f8f86948092d865d0f3bff5ee --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e7fafc8557be797f6d6045ad4f5f29981fd2a49a94e325faaaa05eacbff2bba +size 292814848 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..27d0dc4ed2303d4e54b39f98a6ffce21dc4f58c6 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6366de6d5584f0d172c83cb0f36e28f9a32e9800 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..57fa026f09ae8919a66e13fce96f686bb40eba81 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/trainer_state.json @@ -0,0 +1,12634 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6950075292482335, + "eval_steps": 500, + "global_step": 18000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.096124203008e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..96cb44ab1022bf18ebdce054646fb13968c706ac --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:976497a501428755ea1e47c436600b405bce98388e450590f6bd37673a026424 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..6baba26fef4330afea82a33333165f304157fa4f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff5e78017e5ad55590f094eb01114f15345f02d317ed0a3e717ab0d64cd4eb0d +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..5c94080031253f7346befa04ea6080c1fe949537 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..62a0b29464f27b1b47877b9120d1dfd6a097d572 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..9f85ba597970a4e47ebdb740926d862ffd9c9faf --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/trainer_state.json @@ -0,0 +1,12984 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.7143132939495733, + "eval_steps": 500, + "global_step": 18500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.126572097536e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-18500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e97ac9f48b7c4ada78bc2476334ad79b61dd2da5 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5119ed13e5ed2d2131ea798e8b3f91c1a0d8b96d747f717cfa3d138641f3b0bb +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..b7b35445da185e0a3ed7f609aa8af07039697fe6 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0c1db01bd65b55c6a8f95e1e530a93e63dfdf37ce382ae036799d18e371bbe1 +size 290455552 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..e2e3dbd7cf4c713a498f70c24cc88672a9d62355 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..e473ac408903db0ea7b9c16123e02c9b23e268e8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..64c628d7fe151eb51b9f23aed6667cfe032ca885 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/trainer_state.json @@ -0,0 +1,13334 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.7336190586509131, + "eval_steps": 500, + "global_step": 19000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.157019992064e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..60eff74dd8b886d9c9d25df280cb2870582ba2f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d84ebd13c7821e8ea91508d734c2f19b718290dd16702cce4f44296c18e532a +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..b6b6a941070c0e629ee1bf64f82c5fc0db491d70 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb768b6a734ca53307b6a333c46d7d2911107cd03d595c970c88bcdaf571bbf9 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c76072495b3c6785342efc39c77113f717223edd Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..db399b72df3c6a05de1ce45f85c4de88fffd7052 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..94ecf7c9f640076a26fc777b0cb2750fbcb0eb0f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/trainer_state.json @@ -0,0 +1,13684 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.752924823352253, + "eval_steps": 500, + "global_step": 19500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.187467886592e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-19500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..337a58c049051c08a968fee63254f5feae0fba8c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db6f33ff6afe70379f306301b3875d7bd904aea218d7b6f17d0bf36fdc2ffa0a +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..418dc40dae852c7a87be5cc24a47c2c93327162d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8bdaca7d7e70383691cbcfbb02abc097381d5400e938ee4aa38b7d410af94a61 +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7bd3b78a5e5212d1195c5ba2f6c8f54df05b1f6b Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..aa204ac10abdb639eccce8958bb73182ddd5d843 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..532ab39e3ee7b8b3e8a6dbdb97c0701d2b04d36d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/trainer_state.json @@ -0,0 +1,14034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.7722305880535928, + "eval_steps": 500, + "global_step": 20000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.21791578112e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..48e6507d7c958e15dfd3e3e7611f00906c576724 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b295fae29ae75e5f86b2e233d0f983695e74b3195052b315dcbdbf2ab794c878 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..9818a10977dbb11105b9a16b9f6924ee66f82647 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:397d1ba3ab2ea57ff654af0d0b6ac979a169301b8d02fd9db51771210edc5354 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..de2dfae37d9795d4840bc9aeb9a27098345dc815 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f83a38da040944f69d1c5f2d617490337c679451 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..65969a637cc80fe5370945fba75ffa6a1d272432 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/trainer_state.json @@ -0,0 +1,14384 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.7915363527549326, + "eval_steps": 500, + "global_step": 20500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.248363675648e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-20500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..656de30f5f1e85caf81468e10e02156fa71e0020 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7951275fbdbafba99adcd821adada894ce83168d05a3bab505b897192bc0d74c +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..db3a5e2bfc6d0d70eba9b9579961341d2113355c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d1e9040754d301512fb050adb85f508a43b1d4507ce7dbc3fe142e190d1b766 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..f62748ffeefcc32f035527aece3faf8de51123c6 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..774a563a6e93df1bd4022cb220d59b3bb8e4031e Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..28a838f38b6af06008fb334da4065c18c17a5a74 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/trainer_state.json @@ -0,0 +1,14734 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8108421174562724, + "eval_steps": 500, + "global_step": 21000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.278811570176e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1e4d101b00546bf1f3fbcca07706ad9592c0d87a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13d9f51f19c9ff145e6a154dd1ac182c47800189274275e473e8709854f32c47 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..67100834ba6f13f9c13d785f9a67d9474bd60f2c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:411eb58b1696142ca24d44f8c087b6481e448089209054e8c09ae1928f6b8b02 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..6fe0e83df945852d85af7a54593947ca079ba112 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f526e08ad820793da4e39691a1f6aeb47124f185 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..2f5ff14ec18ad6d026f423d3793878128a84f287 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/trainer_state.json @@ -0,0 +1,15084 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8301478821576123, + "eval_steps": 500, + "global_step": 21500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.309259464704e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-21500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7f07553c992851412ba39dda4ca9a431e9657526 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1fcf68d0dea09f57afe1a6e553e05567f68d20bfd92fd7f89116e6996b448c5 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4b0999f0f3048ab8a95b36ac51ce8732a6c526e0 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:727fd0626abf1d546b780c2ebe2d2932a14a3f7dd35a3aa61c34c95041be8a50 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7940488d917b7bebeaff94bb3f909247efb51f53 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..d79bc5264d890ea93e524f9ce55c0ec3a47700c5 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e383c9575e84e4e177550def5646bdd6978d8db7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/trainer_state.json @@ -0,0 +1,15434 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8494536468589521, + "eval_steps": 500, + "global_step": 22000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.339707359232e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e58a3bd80d26bc61b09b799c60b1a4a84b4741cf --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53b9ed71ab105d3e7b5fbb2b10997ed05888a297600382efbb3503fbdef84848 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..fe345ae891a59742d8de3e15473a8ff41c877310 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a4df4149e54eb70b210f8ec16f87fc441a0cc51454d396daf73fe7e67458447 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..6284fa3f89c74674fda69a0cac13483eb208a2df Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..b59de210815dae1162b734b80e366e66391e8443 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..28024cf390abfa20d5e0dc49cbb5bdf06aec2149 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/trainer_state.json @@ -0,0 +1,15784 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8687594115602919, + "eval_steps": 500, + "global_step": 22500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.37015525376e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-22500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d32d3426a8373cc011df7f732efc3782a2bdaabb --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9abefef7e1c8562827a5ad539fb98d95c521de97bfb19901428a46089018c290 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..a60b64c6f8bf53b82b382752b5e109407e79afc5 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ce42e2bd4305f05366c6ff16d3894a0c97aca302b1a16a004b05936e134bb17 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..44af4761e26b15d8faa92c9e10f6341d2c7f3952 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..969d07e79a1538dd79bf6b583e3abc8610146f07 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..a4cbdd42eb83ed69e3a44fcfb4129063c7c0d2a7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/trainer_state.json @@ -0,0 +1,16134 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.8880651762616317, + "eval_steps": 500, + "global_step": 23000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.400603148288e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a1298e922cd8efcd0db6cd62253ede0ea9536081 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7096faa0b2cf203fda3f779ef9868145252f54071f81ebc555574a0c261e5ac5 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f593ad67e467f39824655249b97fd52864721871 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aca292cb2da3027d8bd233d64e243f134a8ce50d2c3e0aa54972045d9e169d88 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..6615f6b0aa6eac621ac84aa7ac41f5f7bf7fa9d5 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..e798de225d261312be87ea4fd8eb3984c0257ea4 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..3a044615baf0df12683761847143a60c5ef36063 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/trainer_state.json @@ -0,0 +1,16484 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9073709409629716, + "eval_steps": 500, + "global_step": 23500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.431051042816e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-23500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..625a538a7dfb7e4af23e825de49cbc6abafe521e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b812cae17752660163f3afc7effe8944469e053fd4468b33078d394779f615ad +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..da13a99fcf2dbe5531d539b3f90c3cc872f848ee --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f32a8ea507b44bc7ed417b04d45eb3da1e53110df9c97f24727b2fc15a68a1ae +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..9454bb168328e8f438afe8defc180b29536fb4da Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ed4d8b20f8c022f414ea5b360ee0087dccd2a90f Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..693f496b6ea9295827126baf447348e4d45913ef --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/trainer_state.json @@ -0,0 +1,16834 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9266767056643114, + "eval_steps": 500, + "global_step": 24000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.461498937344e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cfef37141998ee194b291e8e14940db5d3eec767 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3ce19ffd4146623ac1ed5b64bed3a34b517ab67e12f0306ea67af017bb830be +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..83eba70bed0f895fea196119164f55c9209950d8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8cc1bcba93537510a0b868855ae0c2a06064b816b8b1e307cc10d1671f2ca1b1 +size 293863424 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..5290fcc464c47d85492d17c2f7c97375af2f634c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f07841aeef1f4c8c5e2f8ba695c447e767aa4379 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..0be1b3ec810b602d0964453921ba02762817447b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/trainer_state.json @@ -0,0 +1,17184 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9459824703656512, + "eval_steps": 500, + "global_step": 24500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.491946831872e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/training_args.bin differ