diff --git a/README.md b/README.md
index 08371015f02382e6fcba318f4aaea54ae52cd3c4..5f4f5625039e17bc7326cc10fbbc7896bb06453b 100644
--- a/README.md
+++ b/README.md
@@ -4,18 +4,6 @@ library_name: peft
## Training procedure
-The following `bitsandbytes` quantization config was used during training:
-- quant_method: bitsandbytes
-- load_in_8bit: True
-- load_in_4bit: False
-- llm_int8_threshold: 6.0
-- llm_int8_skip_modules: None
-- llm_int8_enable_fp32_cpu_offload: False
-- llm_int8_has_fp16_weight: False
-- bnb_4bit_quant_type: fp4
-- bnb_4bit_use_double_quant: False
-- bnb_4bit_compute_dtype: float32
-
The following `bitsandbytes` quantization config was used during training:
- quant_method: bitsandbytes
- load_in_8bit: True
@@ -29,6 +17,5 @@ The following `bitsandbytes` quantization config was used during training:
- bnb_4bit_compute_dtype: float32
### Framework versions
-- PEFT 0.6.0.dev0
- PEFT 0.6.0.dev0
diff --git a/adapter_model.bin b/adapter_model.bin
index f7386f516834b53f43314bd73e439dc33a8c2733..b80def9b80e0fb36ed831dd4c4e1346072c76b09 100644
--- a/adapter_model.bin
+++ b/adapter_model.bin
@@ -1,3 +1,3 @@
version https://git-lfs.github.com/spec/v1
-oid sha256:d17f997023cf8f985a547661db8c0ff488c11f01eaeaca062f1c51fd98fd299d
+oid sha256:a3b13e1a616710fea5e375eaba4c717e346333ccc3d748617ed5bb54d4724e79
size 39409357
diff --git a/checkpoint-1000/README.md b/checkpoint-1000/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-1000/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-1000/adapter_config.json b/checkpoint-1000/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-1000/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-1000/adapter_model.bin b/checkpoint-1000/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..2125dd757e89a68fd696f5498b9e67bbeff3195c
--- /dev/null
+++ b/checkpoint-1000/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:442784f9988ccf98a4c80729f764248b7aece00d5cbe4899fc23821ab36912c5
+size 39409357
diff --git a/checkpoint-1000/optimizer.pt b/checkpoint-1000/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6d0935baa96612fa281284ef995cb71e24126a19
--- /dev/null
+++ b/checkpoint-1000/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a86df865574f7737280e79335d7211d01b31b136c41c37c56c6ff9c2dbed8bdc
+size 78844421
diff --git a/checkpoint-1000/rng_state.pth b/checkpoint-1000/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8576d1dd1c6cb7f5cc1be6601c1c5c20de0a738e
--- /dev/null
+++ b/checkpoint-1000/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b0dfb171332ffdf5185078b01e6c5599c2aa6657e6b96cd2596b1da1f86c2d1f
+size 14575
diff --git a/checkpoint-1000/scheduler.pt b/checkpoint-1000/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..15167329707c7c53bc4a432748ffa98a2919cfb8
--- /dev/null
+++ b/checkpoint-1000/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f90c270e464cbba9ce3ac99c0baf5b79e003ab516981bec7ad9324d178b9d534
+size 627
diff --git a/checkpoint-1000/special_tokens_map.json b/checkpoint-1000/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-1000/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-1000/tokenizer.json b/checkpoint-1000/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-1000/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-1000/tokenizer_config.json b/checkpoint-1000/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-1000/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-1000/trainer_state.json b/checkpoint-1000/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..ae639c72a976dd8b276f71d2a54ce6efbfddd4f5
--- /dev/null
+++ b/checkpoint-1000/trainer_state.json
@@ -0,0 +1,6019 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 25.19685039370079,
+ "eval_steps": 500,
+ "global_step": 1000,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ },
+ {
+ "epoch": 10.1,
+ "learning_rate": 0.00033934294871794873,
+ "loss": 0.7635,
+ "step": 401
+ },
+ {
+ "epoch": 10.13,
+ "learning_rate": 0.00033894230769230773,
+ "loss": 0.7713,
+ "step": 402
+ },
+ {
+ "epoch": 10.15,
+ "learning_rate": 0.0003385416666666667,
+ "loss": 0.827,
+ "step": 403
+ },
+ {
+ "epoch": 10.18,
+ "learning_rate": 0.0003381410256410257,
+ "loss": 0.7844,
+ "step": 404
+ },
+ {
+ "epoch": 10.2,
+ "learning_rate": 0.00033774038461538463,
+ "loss": 0.7839,
+ "step": 405
+ },
+ {
+ "epoch": 10.23,
+ "learning_rate": 0.00033733974358974363,
+ "loss": 0.7702,
+ "step": 406
+ },
+ {
+ "epoch": 10.26,
+ "learning_rate": 0.0003369391025641026,
+ "loss": 0.7855,
+ "step": 407
+ },
+ {
+ "epoch": 10.28,
+ "learning_rate": 0.0003365384615384616,
+ "loss": 0.7669,
+ "step": 408
+ },
+ {
+ "epoch": 10.31,
+ "learning_rate": 0.0003361378205128205,
+ "loss": 0.8054,
+ "step": 409
+ },
+ {
+ "epoch": 10.33,
+ "learning_rate": 0.0003357371794871795,
+ "loss": 0.7792,
+ "step": 410
+ },
+ {
+ "epoch": 10.36,
+ "learning_rate": 0.0003353365384615384,
+ "loss": 0.7986,
+ "step": 411
+ },
+ {
+ "epoch": 10.38,
+ "learning_rate": 0.0003349358974358974,
+ "loss": 0.8269,
+ "step": 412
+ },
+ {
+ "epoch": 10.41,
+ "learning_rate": 0.00033453525641025637,
+ "loss": 0.7923,
+ "step": 413
+ },
+ {
+ "epoch": 10.43,
+ "learning_rate": 0.00033413461538461537,
+ "loss": 0.8242,
+ "step": 414
+ },
+ {
+ "epoch": 10.46,
+ "learning_rate": 0.0003337339743589743,
+ "loss": 0.8496,
+ "step": 415
+ },
+ {
+ "epoch": 10.48,
+ "learning_rate": 0.0003333333333333333,
+ "loss": 0.8236,
+ "step": 416
+ },
+ {
+ "epoch": 10.51,
+ "learning_rate": 0.0003329326923076923,
+ "loss": 0.8685,
+ "step": 417
+ },
+ {
+ "epoch": 10.53,
+ "learning_rate": 0.00033253205128205127,
+ "loss": 0.8542,
+ "step": 418
+ },
+ {
+ "epoch": 10.56,
+ "learning_rate": 0.0003321314102564103,
+ "loss": 0.8185,
+ "step": 419
+ },
+ {
+ "epoch": 10.58,
+ "learning_rate": 0.0003317307692307692,
+ "loss": 0.8477,
+ "step": 420
+ },
+ {
+ "epoch": 10.61,
+ "learning_rate": 0.0003313301282051282,
+ "loss": 0.8372,
+ "step": 421
+ },
+ {
+ "epoch": 10.63,
+ "learning_rate": 0.00033092948717948717,
+ "loss": 0.8731,
+ "step": 422
+ },
+ {
+ "epoch": 10.66,
+ "learning_rate": 0.00033052884615384617,
+ "loss": 0.8322,
+ "step": 423
+ },
+ {
+ "epoch": 10.68,
+ "learning_rate": 0.0003301282051282051,
+ "loss": 0.8514,
+ "step": 424
+ },
+ {
+ "epoch": 10.71,
+ "learning_rate": 0.0003297275641025641,
+ "loss": 0.8564,
+ "step": 425
+ },
+ {
+ "epoch": 10.73,
+ "learning_rate": 0.00032932692307692307,
+ "loss": 0.8278,
+ "step": 426
+ },
+ {
+ "epoch": 10.76,
+ "learning_rate": 0.00032892628205128207,
+ "loss": 0.8408,
+ "step": 427
+ },
+ {
+ "epoch": 10.78,
+ "learning_rate": 0.000328525641025641,
+ "loss": 0.8438,
+ "step": 428
+ },
+ {
+ "epoch": 10.81,
+ "learning_rate": 0.000328125,
+ "loss": 0.8862,
+ "step": 429
+ },
+ {
+ "epoch": 10.83,
+ "learning_rate": 0.00032772435897435897,
+ "loss": 0.8631,
+ "step": 430
+ },
+ {
+ "epoch": 10.86,
+ "learning_rate": 0.00032732371794871797,
+ "loss": 0.883,
+ "step": 431
+ },
+ {
+ "epoch": 10.89,
+ "learning_rate": 0.0003269230769230769,
+ "loss": 0.8644,
+ "step": 432
+ },
+ {
+ "epoch": 10.91,
+ "learning_rate": 0.0003265224358974359,
+ "loss": 0.8907,
+ "step": 433
+ },
+ {
+ "epoch": 10.94,
+ "learning_rate": 0.00032612179487179487,
+ "loss": 0.9141,
+ "step": 434
+ },
+ {
+ "epoch": 10.96,
+ "learning_rate": 0.00032572115384615387,
+ "loss": 0.8877,
+ "step": 435
+ },
+ {
+ "epoch": 10.99,
+ "learning_rate": 0.0003253205128205128,
+ "loss": 0.8521,
+ "step": 436
+ },
+ {
+ "epoch": 11.01,
+ "learning_rate": 0.0003249198717948718,
+ "loss": 0.7842,
+ "step": 437
+ },
+ {
+ "epoch": 11.04,
+ "learning_rate": 0.00032451923076923077,
+ "loss": 0.6516,
+ "step": 438
+ },
+ {
+ "epoch": 11.06,
+ "learning_rate": 0.00032411858974358977,
+ "loss": 0.6549,
+ "step": 439
+ },
+ {
+ "epoch": 11.09,
+ "learning_rate": 0.0003237179487179487,
+ "loss": 0.6736,
+ "step": 440
+ },
+ {
+ "epoch": 11.11,
+ "learning_rate": 0.0003233173076923077,
+ "loss": 0.6625,
+ "step": 441
+ },
+ {
+ "epoch": 11.14,
+ "learning_rate": 0.0003229166666666667,
+ "loss": 0.6466,
+ "step": 442
+ },
+ {
+ "epoch": 11.16,
+ "learning_rate": 0.00032251602564102567,
+ "loss": 0.6362,
+ "step": 443
+ },
+ {
+ "epoch": 11.19,
+ "learning_rate": 0.00032211538461538467,
+ "loss": 0.674,
+ "step": 444
+ },
+ {
+ "epoch": 11.21,
+ "learning_rate": 0.0003217147435897436,
+ "loss": 0.6509,
+ "step": 445
+ },
+ {
+ "epoch": 11.24,
+ "learning_rate": 0.0003213141025641026,
+ "loss": 0.618,
+ "step": 446
+ },
+ {
+ "epoch": 11.26,
+ "learning_rate": 0.00032091346153846156,
+ "loss": 0.6654,
+ "step": 447
+ },
+ {
+ "epoch": 11.29,
+ "learning_rate": 0.00032051282051282057,
+ "loss": 0.6632,
+ "step": 448
+ },
+ {
+ "epoch": 11.31,
+ "learning_rate": 0.00032011217948717946,
+ "loss": 0.6599,
+ "step": 449
+ },
+ {
+ "epoch": 11.34,
+ "learning_rate": 0.00031971153846153846,
+ "loss": 0.6537,
+ "step": 450
+ },
+ {
+ "epoch": 11.36,
+ "learning_rate": 0.0003193108974358974,
+ "loss": 0.678,
+ "step": 451
+ },
+ {
+ "epoch": 11.39,
+ "learning_rate": 0.0003189102564102564,
+ "loss": 0.684,
+ "step": 452
+ },
+ {
+ "epoch": 11.41,
+ "learning_rate": 0.00031850961538461536,
+ "loss": 0.6419,
+ "step": 453
+ },
+ {
+ "epoch": 11.44,
+ "learning_rate": 0.00031810897435897436,
+ "loss": 0.6614,
+ "step": 454
+ },
+ {
+ "epoch": 11.46,
+ "learning_rate": 0.0003177083333333333,
+ "loss": 0.7037,
+ "step": 455
+ },
+ {
+ "epoch": 11.49,
+ "learning_rate": 0.0003173076923076923,
+ "loss": 0.6944,
+ "step": 456
+ },
+ {
+ "epoch": 11.51,
+ "learning_rate": 0.00031690705128205126,
+ "loss": 0.6818,
+ "step": 457
+ },
+ {
+ "epoch": 11.54,
+ "learning_rate": 0.00031650641025641026,
+ "loss": 0.692,
+ "step": 458
+ },
+ {
+ "epoch": 11.57,
+ "learning_rate": 0.0003161057692307692,
+ "loss": 0.7028,
+ "step": 459
+ },
+ {
+ "epoch": 11.59,
+ "learning_rate": 0.0003157051282051282,
+ "loss": 0.7009,
+ "step": 460
+ },
+ {
+ "epoch": 11.62,
+ "learning_rate": 0.00031530448717948716,
+ "loss": 0.7325,
+ "step": 461
+ },
+ {
+ "epoch": 11.64,
+ "learning_rate": 0.00031490384615384616,
+ "loss": 0.6738,
+ "step": 462
+ },
+ {
+ "epoch": 11.67,
+ "learning_rate": 0.0003145032051282051,
+ "loss": 0.7008,
+ "step": 463
+ },
+ {
+ "epoch": 11.69,
+ "learning_rate": 0.0003141025641025641,
+ "loss": 0.6838,
+ "step": 464
+ },
+ {
+ "epoch": 11.72,
+ "learning_rate": 0.00031370192307692306,
+ "loss": 0.726,
+ "step": 465
+ },
+ {
+ "epoch": 11.74,
+ "learning_rate": 0.00031330128205128206,
+ "loss": 0.7167,
+ "step": 466
+ },
+ {
+ "epoch": 11.77,
+ "learning_rate": 0.000312900641025641,
+ "loss": 0.7236,
+ "step": 467
+ },
+ {
+ "epoch": 11.79,
+ "learning_rate": 0.0003125,
+ "loss": 0.7076,
+ "step": 468
+ },
+ {
+ "epoch": 11.82,
+ "learning_rate": 0.000312099358974359,
+ "loss": 0.7415,
+ "step": 469
+ },
+ {
+ "epoch": 11.84,
+ "learning_rate": 0.00031169871794871796,
+ "loss": 0.7313,
+ "step": 470
+ },
+ {
+ "epoch": 11.87,
+ "learning_rate": 0.00031129807692307696,
+ "loss": 0.7403,
+ "step": 471
+ },
+ {
+ "epoch": 11.89,
+ "learning_rate": 0.0003108974358974359,
+ "loss": 0.7145,
+ "step": 472
+ },
+ {
+ "epoch": 11.92,
+ "learning_rate": 0.0003104967948717949,
+ "loss": 0.7532,
+ "step": 473
+ },
+ {
+ "epoch": 11.94,
+ "learning_rate": 0.00031009615384615385,
+ "loss": 0.7326,
+ "step": 474
+ },
+ {
+ "epoch": 11.97,
+ "learning_rate": 0.00030969551282051286,
+ "loss": 0.7699,
+ "step": 475
+ },
+ {
+ "epoch": 11.99,
+ "learning_rate": 0.0003092948717948718,
+ "loss": 0.7251,
+ "step": 476
+ },
+ {
+ "epoch": 12.02,
+ "learning_rate": 0.0003088942307692308,
+ "loss": 0.5773,
+ "step": 477
+ },
+ {
+ "epoch": 12.04,
+ "learning_rate": 0.00030849358974358975,
+ "loss": 0.5288,
+ "step": 478
+ },
+ {
+ "epoch": 12.07,
+ "learning_rate": 0.00030809294871794876,
+ "loss": 0.5312,
+ "step": 479
+ },
+ {
+ "epoch": 12.09,
+ "learning_rate": 0.0003076923076923077,
+ "loss": 0.5445,
+ "step": 480
+ },
+ {
+ "epoch": 12.12,
+ "learning_rate": 0.0003072916666666667,
+ "loss": 0.5218,
+ "step": 481
+ },
+ {
+ "epoch": 12.14,
+ "learning_rate": 0.00030689102564102565,
+ "loss": 0.5311,
+ "step": 482
+ },
+ {
+ "epoch": 12.17,
+ "learning_rate": 0.00030649038461538465,
+ "loss": 0.5541,
+ "step": 483
+ },
+ {
+ "epoch": 12.2,
+ "learning_rate": 0.0003060897435897436,
+ "loss": 0.5242,
+ "step": 484
+ },
+ {
+ "epoch": 12.22,
+ "learning_rate": 0.0003056891025641026,
+ "loss": 0.5302,
+ "step": 485
+ },
+ {
+ "epoch": 12.25,
+ "learning_rate": 0.00030528846153846155,
+ "loss": 0.536,
+ "step": 486
+ },
+ {
+ "epoch": 12.27,
+ "learning_rate": 0.00030488782051282055,
+ "loss": 0.5545,
+ "step": 487
+ },
+ {
+ "epoch": 12.3,
+ "learning_rate": 0.00030448717948717945,
+ "loss": 0.5476,
+ "step": 488
+ },
+ {
+ "epoch": 12.32,
+ "learning_rate": 0.00030408653846153845,
+ "loss": 0.5528,
+ "step": 489
+ },
+ {
+ "epoch": 12.35,
+ "learning_rate": 0.0003036858974358974,
+ "loss": 0.5633,
+ "step": 490
+ },
+ {
+ "epoch": 12.37,
+ "learning_rate": 0.0003032852564102564,
+ "loss": 0.571,
+ "step": 491
+ },
+ {
+ "epoch": 12.4,
+ "learning_rate": 0.00030288461538461535,
+ "loss": 0.5637,
+ "step": 492
+ },
+ {
+ "epoch": 12.42,
+ "learning_rate": 0.00030248397435897435,
+ "loss": 0.5613,
+ "step": 493
+ },
+ {
+ "epoch": 12.45,
+ "learning_rate": 0.0003020833333333333,
+ "loss": 0.5999,
+ "step": 494
+ },
+ {
+ "epoch": 12.47,
+ "learning_rate": 0.0003016826923076923,
+ "loss": 0.5633,
+ "step": 495
+ },
+ {
+ "epoch": 12.5,
+ "learning_rate": 0.0003012820512820513,
+ "loss": 0.5811,
+ "step": 496
+ },
+ {
+ "epoch": 12.52,
+ "learning_rate": 0.00030088141025641025,
+ "loss": 0.5711,
+ "step": 497
+ },
+ {
+ "epoch": 12.55,
+ "learning_rate": 0.00030048076923076925,
+ "loss": 0.5669,
+ "step": 498
+ },
+ {
+ "epoch": 12.57,
+ "learning_rate": 0.0003000801282051282,
+ "loss": 0.5424,
+ "step": 499
+ },
+ {
+ "epoch": 12.6,
+ "learning_rate": 0.0002996794871794872,
+ "loss": 0.5592,
+ "step": 500
+ },
+ {
+ "epoch": 12.62,
+ "learning_rate": 0.00029927884615384614,
+ "loss": 0.5645,
+ "step": 501
+ },
+ {
+ "epoch": 12.65,
+ "learning_rate": 0.00029887820512820515,
+ "loss": 0.6089,
+ "step": 502
+ },
+ {
+ "epoch": 12.67,
+ "learning_rate": 0.0002984775641025641,
+ "loss": 0.57,
+ "step": 503
+ },
+ {
+ "epoch": 12.7,
+ "learning_rate": 0.0002980769230769231,
+ "loss": 0.57,
+ "step": 504
+ },
+ {
+ "epoch": 12.72,
+ "learning_rate": 0.00029767628205128204,
+ "loss": 0.5511,
+ "step": 505
+ },
+ {
+ "epoch": 12.75,
+ "learning_rate": 0.00029727564102564105,
+ "loss": 0.5765,
+ "step": 506
+ },
+ {
+ "epoch": 12.77,
+ "learning_rate": 0.000296875,
+ "loss": 0.5896,
+ "step": 507
+ },
+ {
+ "epoch": 12.8,
+ "learning_rate": 0.000296474358974359,
+ "loss": 0.6329,
+ "step": 508
+ },
+ {
+ "epoch": 12.83,
+ "learning_rate": 0.00029607371794871794,
+ "loss": 0.5682,
+ "step": 509
+ },
+ {
+ "epoch": 12.85,
+ "learning_rate": 0.00029567307692307694,
+ "loss": 0.58,
+ "step": 510
+ },
+ {
+ "epoch": 12.88,
+ "learning_rate": 0.0002952724358974359,
+ "loss": 0.6014,
+ "step": 511
+ },
+ {
+ "epoch": 12.9,
+ "learning_rate": 0.0002948717948717949,
+ "loss": 0.6082,
+ "step": 512
+ },
+ {
+ "epoch": 12.93,
+ "learning_rate": 0.00029447115384615384,
+ "loss": 0.6392,
+ "step": 513
+ },
+ {
+ "epoch": 12.95,
+ "learning_rate": 0.00029407051282051284,
+ "loss": 0.609,
+ "step": 514
+ },
+ {
+ "epoch": 12.98,
+ "learning_rate": 0.0002936698717948718,
+ "loss": 0.6247,
+ "step": 515
+ },
+ {
+ "epoch": 13.0,
+ "learning_rate": 0.0002932692307692308,
+ "loss": 0.5943,
+ "step": 516
+ },
+ {
+ "epoch": 13.03,
+ "learning_rate": 0.00029286858974358974,
+ "loss": 0.4499,
+ "step": 517
+ },
+ {
+ "epoch": 13.05,
+ "learning_rate": 0.00029246794871794874,
+ "loss": 0.4412,
+ "step": 518
+ },
+ {
+ "epoch": 13.08,
+ "learning_rate": 0.0002920673076923077,
+ "loss": 0.4409,
+ "step": 519
+ },
+ {
+ "epoch": 13.1,
+ "learning_rate": 0.0002916666666666667,
+ "loss": 0.4505,
+ "step": 520
+ },
+ {
+ "epoch": 13.13,
+ "learning_rate": 0.0002912660256410257,
+ "loss": 0.4489,
+ "step": 521
+ },
+ {
+ "epoch": 13.15,
+ "learning_rate": 0.00029086538461538464,
+ "loss": 0.4235,
+ "step": 522
+ },
+ {
+ "epoch": 13.18,
+ "learning_rate": 0.00029046474358974364,
+ "loss": 0.4276,
+ "step": 523
+ },
+ {
+ "epoch": 13.2,
+ "learning_rate": 0.0002900641025641026,
+ "loss": 0.4544,
+ "step": 524
+ },
+ {
+ "epoch": 13.23,
+ "learning_rate": 0.0002896634615384616,
+ "loss": 0.4625,
+ "step": 525
+ },
+ {
+ "epoch": 13.25,
+ "learning_rate": 0.00028926282051282054,
+ "loss": 0.4266,
+ "step": 526
+ },
+ {
+ "epoch": 13.28,
+ "learning_rate": 0.0002888621794871795,
+ "loss": 0.4459,
+ "step": 527
+ },
+ {
+ "epoch": 13.3,
+ "learning_rate": 0.00028846153846153843,
+ "loss": 0.4261,
+ "step": 528
+ },
+ {
+ "epoch": 13.33,
+ "learning_rate": 0.00028806089743589744,
+ "loss": 0.4376,
+ "step": 529
+ },
+ {
+ "epoch": 13.35,
+ "learning_rate": 0.0002876602564102564,
+ "loss": 0.4418,
+ "step": 530
+ },
+ {
+ "epoch": 13.38,
+ "learning_rate": 0.0002872596153846154,
+ "loss": 0.4667,
+ "step": 531
+ },
+ {
+ "epoch": 13.4,
+ "learning_rate": 0.00028685897435897433,
+ "loss": 0.4504,
+ "step": 532
+ },
+ {
+ "epoch": 13.43,
+ "learning_rate": 0.00028645833333333333,
+ "loss": 0.4691,
+ "step": 533
+ },
+ {
+ "epoch": 13.46,
+ "learning_rate": 0.0002860576923076923,
+ "loss": 0.4684,
+ "step": 534
+ },
+ {
+ "epoch": 13.48,
+ "learning_rate": 0.0002856570512820513,
+ "loss": 0.467,
+ "step": 535
+ },
+ {
+ "epoch": 13.51,
+ "learning_rate": 0.00028525641025641023,
+ "loss": 0.4716,
+ "step": 536
+ },
+ {
+ "epoch": 13.53,
+ "learning_rate": 0.00028485576923076923,
+ "loss": 0.4619,
+ "step": 537
+ },
+ {
+ "epoch": 13.56,
+ "learning_rate": 0.0002844551282051282,
+ "loss": 0.4775,
+ "step": 538
+ },
+ {
+ "epoch": 13.58,
+ "learning_rate": 0.0002840544871794872,
+ "loss": 0.4568,
+ "step": 539
+ },
+ {
+ "epoch": 13.61,
+ "learning_rate": 0.00028365384615384613,
+ "loss": 0.4763,
+ "step": 540
+ },
+ {
+ "epoch": 13.63,
+ "learning_rate": 0.00028325320512820513,
+ "loss": 0.466,
+ "step": 541
+ },
+ {
+ "epoch": 13.66,
+ "learning_rate": 0.0002828525641025641,
+ "loss": 0.4543,
+ "step": 542
+ },
+ {
+ "epoch": 13.68,
+ "learning_rate": 0.0002824519230769231,
+ "loss": 0.4666,
+ "step": 543
+ },
+ {
+ "epoch": 13.71,
+ "learning_rate": 0.00028205128205128203,
+ "loss": 0.489,
+ "step": 544
+ },
+ {
+ "epoch": 13.73,
+ "learning_rate": 0.00028165064102564103,
+ "loss": 0.491,
+ "step": 545
+ },
+ {
+ "epoch": 13.76,
+ "learning_rate": 0.00028125000000000003,
+ "loss": 0.4847,
+ "step": 546
+ },
+ {
+ "epoch": 13.78,
+ "learning_rate": 0.000280849358974359,
+ "loss": 0.4899,
+ "step": 547
+ },
+ {
+ "epoch": 13.81,
+ "learning_rate": 0.000280448717948718,
+ "loss": 0.5169,
+ "step": 548
+ },
+ {
+ "epoch": 13.83,
+ "learning_rate": 0.00028004807692307693,
+ "loss": 0.4968,
+ "step": 549
+ },
+ {
+ "epoch": 13.86,
+ "learning_rate": 0.00027964743589743593,
+ "loss": 0.4774,
+ "step": 550
+ },
+ {
+ "epoch": 13.88,
+ "learning_rate": 0.0002792467948717949,
+ "loss": 0.5105,
+ "step": 551
+ },
+ {
+ "epoch": 13.91,
+ "learning_rate": 0.0002788461538461539,
+ "loss": 0.4997,
+ "step": 552
+ },
+ {
+ "epoch": 13.93,
+ "learning_rate": 0.00027844551282051283,
+ "loss": 0.5088,
+ "step": 553
+ },
+ {
+ "epoch": 13.96,
+ "learning_rate": 0.00027804487179487183,
+ "loss": 0.5091,
+ "step": 554
+ },
+ {
+ "epoch": 13.98,
+ "learning_rate": 0.0002776442307692308,
+ "loss": 0.532,
+ "step": 555
+ },
+ {
+ "epoch": 14.01,
+ "learning_rate": 0.0002772435897435898,
+ "loss": 0.4487,
+ "step": 556
+ },
+ {
+ "epoch": 14.03,
+ "learning_rate": 0.00027684294871794873,
+ "loss": 0.3525,
+ "step": 557
+ },
+ {
+ "epoch": 14.06,
+ "learning_rate": 0.00027644230769230773,
+ "loss": 0.3715,
+ "step": 558
+ },
+ {
+ "epoch": 14.09,
+ "learning_rate": 0.0002760416666666667,
+ "loss": 0.3547,
+ "step": 559
+ },
+ {
+ "epoch": 14.11,
+ "learning_rate": 0.0002756410256410257,
+ "loss": 0.3862,
+ "step": 560
+ },
+ {
+ "epoch": 14.14,
+ "learning_rate": 0.0002752403846153846,
+ "loss": 0.3514,
+ "step": 561
+ },
+ {
+ "epoch": 14.16,
+ "learning_rate": 0.00027483974358974363,
+ "loss": 0.362,
+ "step": 562
+ },
+ {
+ "epoch": 14.19,
+ "learning_rate": 0.0002744391025641026,
+ "loss": 0.3457,
+ "step": 563
+ },
+ {
+ "epoch": 14.21,
+ "learning_rate": 0.0002740384615384616,
+ "loss": 0.3662,
+ "step": 564
+ },
+ {
+ "epoch": 14.24,
+ "learning_rate": 0.0002736378205128205,
+ "loss": 0.3746,
+ "step": 565
+ },
+ {
+ "epoch": 14.26,
+ "learning_rate": 0.00027323717948717947,
+ "loss": 0.3388,
+ "step": 566
+ },
+ {
+ "epoch": 14.29,
+ "learning_rate": 0.0002728365384615384,
+ "loss": 0.3824,
+ "step": 567
+ },
+ {
+ "epoch": 14.31,
+ "learning_rate": 0.0002724358974358974,
+ "loss": 0.3584,
+ "step": 568
+ },
+ {
+ "epoch": 14.34,
+ "learning_rate": 0.00027203525641025637,
+ "loss": 0.3544,
+ "step": 569
+ },
+ {
+ "epoch": 14.36,
+ "learning_rate": 0.00027163461538461537,
+ "loss": 0.3721,
+ "step": 570
+ },
+ {
+ "epoch": 14.39,
+ "learning_rate": 0.0002712339743589743,
+ "loss": 0.3769,
+ "step": 571
+ },
+ {
+ "epoch": 14.41,
+ "learning_rate": 0.0002708333333333333,
+ "loss": 0.3872,
+ "step": 572
+ },
+ {
+ "epoch": 14.44,
+ "learning_rate": 0.0002704326923076923,
+ "loss": 0.363,
+ "step": 573
+ },
+ {
+ "epoch": 14.46,
+ "learning_rate": 0.00027003205128205127,
+ "loss": 0.3842,
+ "step": 574
+ },
+ {
+ "epoch": 14.49,
+ "learning_rate": 0.00026963141025641027,
+ "loss": 0.3737,
+ "step": 575
+ },
+ {
+ "epoch": 14.51,
+ "learning_rate": 0.0002692307692307692,
+ "loss": 0.3981,
+ "step": 576
+ },
+ {
+ "epoch": 14.54,
+ "learning_rate": 0.0002688301282051282,
+ "loss": 0.3656,
+ "step": 577
+ },
+ {
+ "epoch": 14.56,
+ "learning_rate": 0.00026842948717948717,
+ "loss": 0.3852,
+ "step": 578
+ },
+ {
+ "epoch": 14.59,
+ "learning_rate": 0.00026802884615384617,
+ "loss": 0.3865,
+ "step": 579
+ },
+ {
+ "epoch": 14.61,
+ "learning_rate": 0.0002676282051282051,
+ "loss": 0.3747,
+ "step": 580
+ },
+ {
+ "epoch": 14.64,
+ "learning_rate": 0.0002672275641025641,
+ "loss": 0.3729,
+ "step": 581
+ },
+ {
+ "epoch": 14.66,
+ "learning_rate": 0.00026682692307692307,
+ "loss": 0.3785,
+ "step": 582
+ },
+ {
+ "epoch": 14.69,
+ "learning_rate": 0.00026642628205128207,
+ "loss": 0.4018,
+ "step": 583
+ },
+ {
+ "epoch": 14.71,
+ "learning_rate": 0.000266025641025641,
+ "loss": 0.3854,
+ "step": 584
+ },
+ {
+ "epoch": 14.74,
+ "learning_rate": 0.000265625,
+ "loss": 0.4161,
+ "step": 585
+ },
+ {
+ "epoch": 14.77,
+ "learning_rate": 0.00026522435897435897,
+ "loss": 0.4093,
+ "step": 586
+ },
+ {
+ "epoch": 14.79,
+ "learning_rate": 0.00026482371794871797,
+ "loss": 0.4126,
+ "step": 587
+ },
+ {
+ "epoch": 14.82,
+ "learning_rate": 0.0002644230769230769,
+ "loss": 0.3977,
+ "step": 588
+ },
+ {
+ "epoch": 14.84,
+ "learning_rate": 0.0002640224358974359,
+ "loss": 0.4286,
+ "step": 589
+ },
+ {
+ "epoch": 14.87,
+ "learning_rate": 0.00026362179487179487,
+ "loss": 0.4371,
+ "step": 590
+ },
+ {
+ "epoch": 14.89,
+ "learning_rate": 0.00026322115384615387,
+ "loss": 0.4066,
+ "step": 591
+ },
+ {
+ "epoch": 14.92,
+ "learning_rate": 0.0002628205128205128,
+ "loss": 0.433,
+ "step": 592
+ },
+ {
+ "epoch": 14.94,
+ "learning_rate": 0.0002624198717948718,
+ "loss": 0.4361,
+ "step": 593
+ },
+ {
+ "epoch": 14.97,
+ "learning_rate": 0.00026201923076923076,
+ "loss": 0.4276,
+ "step": 594
+ },
+ {
+ "epoch": 14.99,
+ "learning_rate": 0.00026161858974358977,
+ "loss": 0.4401,
+ "step": 595
+ },
+ {
+ "epoch": 15.02,
+ "learning_rate": 0.0002612179487179487,
+ "loss": 0.3514,
+ "step": 596
+ },
+ {
+ "epoch": 15.04,
+ "learning_rate": 0.0002608173076923077,
+ "loss": 0.306,
+ "step": 597
+ },
+ {
+ "epoch": 15.07,
+ "learning_rate": 0.0002604166666666667,
+ "loss": 0.2862,
+ "step": 598
+ },
+ {
+ "epoch": 15.09,
+ "learning_rate": 0.00026001602564102566,
+ "loss": 0.3042,
+ "step": 599
+ },
+ {
+ "epoch": 15.12,
+ "learning_rate": 0.00025961538461538467,
+ "loss": 0.2978,
+ "step": 600
+ },
+ {
+ "epoch": 15.14,
+ "learning_rate": 0.0002592147435897436,
+ "loss": 0.2975,
+ "step": 601
+ },
+ {
+ "epoch": 15.17,
+ "learning_rate": 0.0002588141025641026,
+ "loss": 0.3172,
+ "step": 602
+ },
+ {
+ "epoch": 15.19,
+ "learning_rate": 0.00025841346153846156,
+ "loss": 0.3046,
+ "step": 603
+ },
+ {
+ "epoch": 15.22,
+ "learning_rate": 0.00025801282051282057,
+ "loss": 0.3008,
+ "step": 604
+ },
+ {
+ "epoch": 15.24,
+ "learning_rate": 0.00025761217948717946,
+ "loss": 0.2956,
+ "step": 605
+ },
+ {
+ "epoch": 15.27,
+ "learning_rate": 0.00025721153846153846,
+ "loss": 0.3137,
+ "step": 606
+ },
+ {
+ "epoch": 15.29,
+ "learning_rate": 0.0002568108974358974,
+ "loss": 0.2961,
+ "step": 607
+ },
+ {
+ "epoch": 15.32,
+ "learning_rate": 0.0002564102564102564,
+ "loss": 0.3059,
+ "step": 608
+ },
+ {
+ "epoch": 15.34,
+ "learning_rate": 0.00025600961538461536,
+ "loss": 0.3077,
+ "step": 609
+ },
+ {
+ "epoch": 15.37,
+ "learning_rate": 0.00025560897435897436,
+ "loss": 0.3189,
+ "step": 610
+ },
+ {
+ "epoch": 15.4,
+ "learning_rate": 0.0002552083333333333,
+ "loss": 0.309,
+ "step": 611
+ },
+ {
+ "epoch": 15.42,
+ "learning_rate": 0.0002548076923076923,
+ "loss": 0.3324,
+ "step": 612
+ },
+ {
+ "epoch": 15.45,
+ "learning_rate": 0.00025440705128205126,
+ "loss": 0.3159,
+ "step": 613
+ },
+ {
+ "epoch": 15.47,
+ "learning_rate": 0.00025400641025641026,
+ "loss": 0.3188,
+ "step": 614
+ },
+ {
+ "epoch": 15.5,
+ "learning_rate": 0.0002536057692307692,
+ "loss": 0.3174,
+ "step": 615
+ },
+ {
+ "epoch": 15.52,
+ "learning_rate": 0.0002532051282051282,
+ "loss": 0.3074,
+ "step": 616
+ },
+ {
+ "epoch": 15.55,
+ "learning_rate": 0.00025280448717948716,
+ "loss": 0.3145,
+ "step": 617
+ },
+ {
+ "epoch": 15.57,
+ "learning_rate": 0.00025240384615384616,
+ "loss": 0.317,
+ "step": 618
+ },
+ {
+ "epoch": 15.6,
+ "learning_rate": 0.0002520032051282051,
+ "loss": 0.318,
+ "step": 619
+ },
+ {
+ "epoch": 15.62,
+ "learning_rate": 0.0002516025641025641,
+ "loss": 0.3236,
+ "step": 620
+ },
+ {
+ "epoch": 15.65,
+ "learning_rate": 0.00025120192307692305,
+ "loss": 0.3282,
+ "step": 621
+ },
+ {
+ "epoch": 15.67,
+ "learning_rate": 0.00025080128205128206,
+ "loss": 0.3394,
+ "step": 622
+ },
+ {
+ "epoch": 15.7,
+ "learning_rate": 0.000250400641025641,
+ "loss": 0.3278,
+ "step": 623
+ },
+ {
+ "epoch": 15.72,
+ "learning_rate": 0.00025,
+ "loss": 0.3255,
+ "step": 624
+ },
+ {
+ "epoch": 15.75,
+ "learning_rate": 0.000249599358974359,
+ "loss": 0.3256,
+ "step": 625
+ },
+ {
+ "epoch": 15.77,
+ "learning_rate": 0.00024919871794871795,
+ "loss": 0.3379,
+ "step": 626
+ },
+ {
+ "epoch": 15.8,
+ "learning_rate": 0.00024879807692307696,
+ "loss": 0.3384,
+ "step": 627
+ },
+ {
+ "epoch": 15.82,
+ "learning_rate": 0.0002483974358974359,
+ "loss": 0.3483,
+ "step": 628
+ },
+ {
+ "epoch": 15.85,
+ "learning_rate": 0.0002479967948717949,
+ "loss": 0.3478,
+ "step": 629
+ },
+ {
+ "epoch": 15.87,
+ "learning_rate": 0.00024759615384615385,
+ "loss": 0.337,
+ "step": 630
+ },
+ {
+ "epoch": 15.9,
+ "learning_rate": 0.00024719551282051286,
+ "loss": 0.3287,
+ "step": 631
+ },
+ {
+ "epoch": 15.92,
+ "learning_rate": 0.0002467948717948718,
+ "loss": 0.3478,
+ "step": 632
+ },
+ {
+ "epoch": 15.95,
+ "learning_rate": 0.0002463942307692308,
+ "loss": 0.3567,
+ "step": 633
+ },
+ {
+ "epoch": 15.97,
+ "learning_rate": 0.00024599358974358975,
+ "loss": 0.3359,
+ "step": 634
+ },
+ {
+ "epoch": 16.0,
+ "learning_rate": 0.0002455929487179487,
+ "loss": 0.3492,
+ "step": 635
+ },
+ {
+ "epoch": 16.03,
+ "learning_rate": 0.0002451923076923077,
+ "loss": 0.2497,
+ "step": 636
+ },
+ {
+ "epoch": 16.05,
+ "learning_rate": 0.00024479166666666665,
+ "loss": 0.2416,
+ "step": 637
+ },
+ {
+ "epoch": 16.08,
+ "learning_rate": 0.00024439102564102565,
+ "loss": 0.2576,
+ "step": 638
+ },
+ {
+ "epoch": 16.1,
+ "learning_rate": 0.00024399038461538463,
+ "loss": 0.243,
+ "step": 639
+ },
+ {
+ "epoch": 16.13,
+ "learning_rate": 0.0002435897435897436,
+ "loss": 0.2527,
+ "step": 640
+ },
+ {
+ "epoch": 16.15,
+ "learning_rate": 0.00024318910256410258,
+ "loss": 0.2606,
+ "step": 641
+ },
+ {
+ "epoch": 16.18,
+ "learning_rate": 0.00024278846153846155,
+ "loss": 0.2555,
+ "step": 642
+ },
+ {
+ "epoch": 16.2,
+ "learning_rate": 0.00024238782051282052,
+ "loss": 0.2581,
+ "step": 643
+ },
+ {
+ "epoch": 16.23,
+ "learning_rate": 0.0002419871794871795,
+ "loss": 0.2508,
+ "step": 644
+ },
+ {
+ "epoch": 16.25,
+ "learning_rate": 0.00024158653846153847,
+ "loss": 0.281,
+ "step": 645
+ },
+ {
+ "epoch": 16.28,
+ "learning_rate": 0.00024118589743589745,
+ "loss": 0.2626,
+ "step": 646
+ },
+ {
+ "epoch": 16.3,
+ "learning_rate": 0.00024078525641025642,
+ "loss": 0.2668,
+ "step": 647
+ },
+ {
+ "epoch": 16.33,
+ "learning_rate": 0.0002403846153846154,
+ "loss": 0.2675,
+ "step": 648
+ },
+ {
+ "epoch": 16.35,
+ "learning_rate": 0.00023998397435897437,
+ "loss": 0.2491,
+ "step": 649
+ },
+ {
+ "epoch": 16.38,
+ "learning_rate": 0.00023958333333333335,
+ "loss": 0.2459,
+ "step": 650
+ },
+ {
+ "epoch": 16.4,
+ "learning_rate": 0.00023918269230769232,
+ "loss": 0.2515,
+ "step": 651
+ },
+ {
+ "epoch": 16.43,
+ "learning_rate": 0.0002387820512820513,
+ "loss": 0.2612,
+ "step": 652
+ },
+ {
+ "epoch": 16.45,
+ "learning_rate": 0.00023838141025641027,
+ "loss": 0.2612,
+ "step": 653
+ },
+ {
+ "epoch": 16.48,
+ "learning_rate": 0.00023798076923076922,
+ "loss": 0.265,
+ "step": 654
+ },
+ {
+ "epoch": 16.5,
+ "learning_rate": 0.0002375801282051282,
+ "loss": 0.2634,
+ "step": 655
+ },
+ {
+ "epoch": 16.53,
+ "learning_rate": 0.00023717948717948717,
+ "loss": 0.2839,
+ "step": 656
+ },
+ {
+ "epoch": 16.55,
+ "learning_rate": 0.00023677884615384614,
+ "loss": 0.2595,
+ "step": 657
+ },
+ {
+ "epoch": 16.58,
+ "learning_rate": 0.00023637820512820512,
+ "loss": 0.2764,
+ "step": 658
+ },
+ {
+ "epoch": 16.6,
+ "learning_rate": 0.0002359775641025641,
+ "loss": 0.2553,
+ "step": 659
+ },
+ {
+ "epoch": 16.63,
+ "learning_rate": 0.00023557692307692307,
+ "loss": 0.2716,
+ "step": 660
+ },
+ {
+ "epoch": 16.66,
+ "learning_rate": 0.00023517628205128204,
+ "loss": 0.289,
+ "step": 661
+ },
+ {
+ "epoch": 16.68,
+ "learning_rate": 0.00023477564102564102,
+ "loss": 0.2483,
+ "step": 662
+ },
+ {
+ "epoch": 16.71,
+ "learning_rate": 0.000234375,
+ "loss": 0.269,
+ "step": 663
+ },
+ {
+ "epoch": 16.73,
+ "learning_rate": 0.000233974358974359,
+ "loss": 0.2738,
+ "step": 664
+ },
+ {
+ "epoch": 16.76,
+ "learning_rate": 0.00023357371794871797,
+ "loss": 0.2646,
+ "step": 665
+ },
+ {
+ "epoch": 16.78,
+ "learning_rate": 0.00023317307692307694,
+ "loss": 0.2793,
+ "step": 666
+ },
+ {
+ "epoch": 16.81,
+ "learning_rate": 0.00023277243589743592,
+ "loss": 0.2666,
+ "step": 667
+ },
+ {
+ "epoch": 16.83,
+ "learning_rate": 0.0002323717948717949,
+ "loss": 0.2799,
+ "step": 668
+ },
+ {
+ "epoch": 16.86,
+ "learning_rate": 0.00023197115384615387,
+ "loss": 0.2677,
+ "step": 669
+ },
+ {
+ "epoch": 16.88,
+ "learning_rate": 0.00023157051282051284,
+ "loss": 0.2889,
+ "step": 670
+ },
+ {
+ "epoch": 16.91,
+ "learning_rate": 0.00023116987179487182,
+ "loss": 0.2869,
+ "step": 671
+ },
+ {
+ "epoch": 16.93,
+ "learning_rate": 0.0002307692307692308,
+ "loss": 0.2814,
+ "step": 672
+ },
+ {
+ "epoch": 16.96,
+ "learning_rate": 0.00023036858974358974,
+ "loss": 0.2881,
+ "step": 673
+ },
+ {
+ "epoch": 16.98,
+ "learning_rate": 0.0002299679487179487,
+ "loss": 0.2777,
+ "step": 674
+ },
+ {
+ "epoch": 17.01,
+ "learning_rate": 0.0002295673076923077,
+ "loss": 0.2713,
+ "step": 675
+ },
+ {
+ "epoch": 17.03,
+ "learning_rate": 0.00022916666666666666,
+ "loss": 0.2176,
+ "step": 676
+ },
+ {
+ "epoch": 17.06,
+ "learning_rate": 0.00022876602564102564,
+ "loss": 0.2032,
+ "step": 677
+ },
+ {
+ "epoch": 17.08,
+ "learning_rate": 0.0002283653846153846,
+ "loss": 0.2207,
+ "step": 678
+ },
+ {
+ "epoch": 17.11,
+ "learning_rate": 0.0002279647435897436,
+ "loss": 0.2183,
+ "step": 679
+ },
+ {
+ "epoch": 17.13,
+ "learning_rate": 0.00022756410256410256,
+ "loss": 0.2238,
+ "step": 680
+ },
+ {
+ "epoch": 17.16,
+ "learning_rate": 0.00022716346153846154,
+ "loss": 0.2126,
+ "step": 681
+ },
+ {
+ "epoch": 17.18,
+ "learning_rate": 0.0002267628205128205,
+ "loss": 0.2104,
+ "step": 682
+ },
+ {
+ "epoch": 17.21,
+ "learning_rate": 0.00022636217948717949,
+ "loss": 0.2151,
+ "step": 683
+ },
+ {
+ "epoch": 17.23,
+ "learning_rate": 0.00022596153846153846,
+ "loss": 0.2093,
+ "step": 684
+ },
+ {
+ "epoch": 17.26,
+ "learning_rate": 0.00022556089743589743,
+ "loss": 0.2308,
+ "step": 685
+ },
+ {
+ "epoch": 17.29,
+ "learning_rate": 0.0002251602564102564,
+ "loss": 0.2276,
+ "step": 686
+ },
+ {
+ "epoch": 17.31,
+ "learning_rate": 0.00022475961538461538,
+ "loss": 0.2328,
+ "step": 687
+ },
+ {
+ "epoch": 17.34,
+ "learning_rate": 0.00022435897435897436,
+ "loss": 0.2076,
+ "step": 688
+ },
+ {
+ "epoch": 17.36,
+ "learning_rate": 0.00022395833333333336,
+ "loss": 0.2125,
+ "step": 689
+ },
+ {
+ "epoch": 17.39,
+ "learning_rate": 0.00022355769230769234,
+ "loss": 0.2181,
+ "step": 690
+ },
+ {
+ "epoch": 17.41,
+ "learning_rate": 0.0002231570512820513,
+ "loss": 0.2153,
+ "step": 691
+ },
+ {
+ "epoch": 17.44,
+ "learning_rate": 0.00022275641025641028,
+ "loss": 0.2122,
+ "step": 692
+ },
+ {
+ "epoch": 17.46,
+ "learning_rate": 0.00022235576923076923,
+ "loss": 0.2262,
+ "step": 693
+ },
+ {
+ "epoch": 17.49,
+ "learning_rate": 0.0002219551282051282,
+ "loss": 0.2206,
+ "step": 694
+ },
+ {
+ "epoch": 17.51,
+ "learning_rate": 0.00022155448717948718,
+ "loss": 0.2223,
+ "step": 695
+ },
+ {
+ "epoch": 17.54,
+ "learning_rate": 0.00022115384615384616,
+ "loss": 0.2306,
+ "step": 696
+ },
+ {
+ "epoch": 17.56,
+ "learning_rate": 0.00022075320512820513,
+ "loss": 0.213,
+ "step": 697
+ },
+ {
+ "epoch": 17.59,
+ "learning_rate": 0.0002203525641025641,
+ "loss": 0.2198,
+ "step": 698
+ },
+ {
+ "epoch": 17.61,
+ "learning_rate": 0.00021995192307692308,
+ "loss": 0.2299,
+ "step": 699
+ },
+ {
+ "epoch": 17.64,
+ "learning_rate": 0.00021955128205128206,
+ "loss": 0.2213,
+ "step": 700
+ },
+ {
+ "epoch": 17.66,
+ "learning_rate": 0.00021915064102564103,
+ "loss": 0.2183,
+ "step": 701
+ },
+ {
+ "epoch": 17.69,
+ "learning_rate": 0.00021875,
+ "loss": 0.2301,
+ "step": 702
+ },
+ {
+ "epoch": 17.71,
+ "learning_rate": 0.00021834935897435898,
+ "loss": 0.2348,
+ "step": 703
+ },
+ {
+ "epoch": 17.74,
+ "learning_rate": 0.00021794871794871795,
+ "loss": 0.229,
+ "step": 704
+ },
+ {
+ "epoch": 17.76,
+ "learning_rate": 0.00021754807692307693,
+ "loss": 0.2397,
+ "step": 705
+ },
+ {
+ "epoch": 17.79,
+ "learning_rate": 0.0002171474358974359,
+ "loss": 0.2254,
+ "step": 706
+ },
+ {
+ "epoch": 17.81,
+ "learning_rate": 0.00021674679487179488,
+ "loss": 0.2306,
+ "step": 707
+ },
+ {
+ "epoch": 17.84,
+ "learning_rate": 0.00021634615384615385,
+ "loss": 0.2388,
+ "step": 708
+ },
+ {
+ "epoch": 17.86,
+ "learning_rate": 0.00021594551282051283,
+ "loss": 0.2374,
+ "step": 709
+ },
+ {
+ "epoch": 17.89,
+ "learning_rate": 0.0002155448717948718,
+ "loss": 0.2412,
+ "step": 710
+ },
+ {
+ "epoch": 17.91,
+ "learning_rate": 0.00021514423076923078,
+ "loss": 0.2203,
+ "step": 711
+ },
+ {
+ "epoch": 17.94,
+ "learning_rate": 0.00021474358974358972,
+ "loss": 0.2359,
+ "step": 712
+ },
+ {
+ "epoch": 17.97,
+ "learning_rate": 0.0002143429487179487,
+ "loss": 0.2343,
+ "step": 713
+ },
+ {
+ "epoch": 17.99,
+ "learning_rate": 0.00021394230769230767,
+ "loss": 0.2397,
+ "step": 714
+ },
+ {
+ "epoch": 18.02,
+ "learning_rate": 0.00021354166666666665,
+ "loss": 0.2055,
+ "step": 715
+ },
+ {
+ "epoch": 18.04,
+ "learning_rate": 0.00021314102564102565,
+ "loss": 0.1723,
+ "step": 716
+ },
+ {
+ "epoch": 18.07,
+ "learning_rate": 0.00021274038461538463,
+ "loss": 0.1782,
+ "step": 717
+ },
+ {
+ "epoch": 18.09,
+ "learning_rate": 0.0002123397435897436,
+ "loss": 0.1813,
+ "step": 718
+ },
+ {
+ "epoch": 18.12,
+ "learning_rate": 0.00021193910256410257,
+ "loss": 0.1834,
+ "step": 719
+ },
+ {
+ "epoch": 18.14,
+ "learning_rate": 0.00021153846153846155,
+ "loss": 0.1759,
+ "step": 720
+ },
+ {
+ "epoch": 18.17,
+ "learning_rate": 0.00021113782051282052,
+ "loss": 0.1855,
+ "step": 721
+ },
+ {
+ "epoch": 18.19,
+ "learning_rate": 0.0002107371794871795,
+ "loss": 0.1756,
+ "step": 722
+ },
+ {
+ "epoch": 18.22,
+ "learning_rate": 0.00021033653846153847,
+ "loss": 0.1783,
+ "step": 723
+ },
+ {
+ "epoch": 18.24,
+ "learning_rate": 0.00020993589743589745,
+ "loss": 0.1899,
+ "step": 724
+ },
+ {
+ "epoch": 18.27,
+ "learning_rate": 0.00020953525641025642,
+ "loss": 0.1817,
+ "step": 725
+ },
+ {
+ "epoch": 18.29,
+ "learning_rate": 0.0002091346153846154,
+ "loss": 0.1856,
+ "step": 726
+ },
+ {
+ "epoch": 18.32,
+ "learning_rate": 0.00020873397435897437,
+ "loss": 0.1849,
+ "step": 727
+ },
+ {
+ "epoch": 18.34,
+ "learning_rate": 0.00020833333333333335,
+ "loss": 0.1758,
+ "step": 728
+ },
+ {
+ "epoch": 18.37,
+ "learning_rate": 0.00020793269230769232,
+ "loss": 0.1907,
+ "step": 729
+ },
+ {
+ "epoch": 18.39,
+ "learning_rate": 0.0002075320512820513,
+ "loss": 0.1836,
+ "step": 730
+ },
+ {
+ "epoch": 18.42,
+ "learning_rate": 0.00020713141025641027,
+ "loss": 0.1828,
+ "step": 731
+ },
+ {
+ "epoch": 18.44,
+ "learning_rate": 0.00020673076923076922,
+ "loss": 0.1916,
+ "step": 732
+ },
+ {
+ "epoch": 18.47,
+ "learning_rate": 0.0002063301282051282,
+ "loss": 0.1849,
+ "step": 733
+ },
+ {
+ "epoch": 18.49,
+ "learning_rate": 0.00020592948717948717,
+ "loss": 0.1929,
+ "step": 734
+ },
+ {
+ "epoch": 18.52,
+ "learning_rate": 0.00020552884615384614,
+ "loss": 0.1779,
+ "step": 735
+ },
+ {
+ "epoch": 18.54,
+ "learning_rate": 0.00020512820512820512,
+ "loss": 0.1805,
+ "step": 736
+ },
+ {
+ "epoch": 18.57,
+ "learning_rate": 0.0002047275641025641,
+ "loss": 0.1891,
+ "step": 737
+ },
+ {
+ "epoch": 18.6,
+ "learning_rate": 0.00020432692307692307,
+ "loss": 0.1932,
+ "step": 738
+ },
+ {
+ "epoch": 18.62,
+ "learning_rate": 0.00020392628205128204,
+ "loss": 0.1923,
+ "step": 739
+ },
+ {
+ "epoch": 18.65,
+ "learning_rate": 0.00020352564102564102,
+ "loss": 0.1882,
+ "step": 740
+ },
+ {
+ "epoch": 18.67,
+ "learning_rate": 0.00020312500000000002,
+ "loss": 0.2012,
+ "step": 741
+ },
+ {
+ "epoch": 18.7,
+ "learning_rate": 0.000202724358974359,
+ "loss": 0.1963,
+ "step": 742
+ },
+ {
+ "epoch": 18.72,
+ "learning_rate": 0.00020232371794871797,
+ "loss": 0.202,
+ "step": 743
+ },
+ {
+ "epoch": 18.75,
+ "learning_rate": 0.00020192307692307694,
+ "loss": 0.1996,
+ "step": 744
+ },
+ {
+ "epoch": 18.77,
+ "learning_rate": 0.00020152243589743592,
+ "loss": 0.2056,
+ "step": 745
+ },
+ {
+ "epoch": 18.8,
+ "learning_rate": 0.0002011217948717949,
+ "loss": 0.1946,
+ "step": 746
+ },
+ {
+ "epoch": 18.82,
+ "learning_rate": 0.00020072115384615387,
+ "loss": 0.1954,
+ "step": 747
+ },
+ {
+ "epoch": 18.85,
+ "learning_rate": 0.00020032051282051284,
+ "loss": 0.1966,
+ "step": 748
+ },
+ {
+ "epoch": 18.87,
+ "learning_rate": 0.00019991987179487182,
+ "loss": 0.2077,
+ "step": 749
+ },
+ {
+ "epoch": 18.9,
+ "learning_rate": 0.0001995192307692308,
+ "loss": 0.2,
+ "step": 750
+ },
+ {
+ "epoch": 18.92,
+ "learning_rate": 0.00019911858974358974,
+ "loss": 0.2092,
+ "step": 751
+ },
+ {
+ "epoch": 18.95,
+ "learning_rate": 0.0001987179487179487,
+ "loss": 0.2023,
+ "step": 752
+ },
+ {
+ "epoch": 18.97,
+ "learning_rate": 0.0001983173076923077,
+ "loss": 0.2056,
+ "step": 753
+ },
+ {
+ "epoch": 19.0,
+ "learning_rate": 0.00019791666666666666,
+ "loss": 0.2059,
+ "step": 754
+ },
+ {
+ "epoch": 19.02,
+ "learning_rate": 0.00019751602564102564,
+ "loss": 0.1581,
+ "step": 755
+ },
+ {
+ "epoch": 19.05,
+ "learning_rate": 0.0001971153846153846,
+ "loss": 0.155,
+ "step": 756
+ },
+ {
+ "epoch": 19.07,
+ "learning_rate": 0.00019671474358974359,
+ "loss": 0.156,
+ "step": 757
+ },
+ {
+ "epoch": 19.1,
+ "learning_rate": 0.00019631410256410256,
+ "loss": 0.161,
+ "step": 758
+ },
+ {
+ "epoch": 19.12,
+ "learning_rate": 0.00019591346153846154,
+ "loss": 0.1608,
+ "step": 759
+ },
+ {
+ "epoch": 19.15,
+ "learning_rate": 0.0001955128205128205,
+ "loss": 0.1576,
+ "step": 760
+ },
+ {
+ "epoch": 19.17,
+ "learning_rate": 0.00019511217948717948,
+ "loss": 0.1539,
+ "step": 761
+ },
+ {
+ "epoch": 19.2,
+ "learning_rate": 0.00019471153846153846,
+ "loss": 0.1549,
+ "step": 762
+ },
+ {
+ "epoch": 19.23,
+ "learning_rate": 0.00019431089743589743,
+ "loss": 0.1593,
+ "step": 763
+ },
+ {
+ "epoch": 19.25,
+ "learning_rate": 0.0001939102564102564,
+ "loss": 0.162,
+ "step": 764
+ },
+ {
+ "epoch": 19.28,
+ "learning_rate": 0.00019350961538461538,
+ "loss": 0.1591,
+ "step": 765
+ },
+ {
+ "epoch": 19.3,
+ "learning_rate": 0.00019310897435897436,
+ "loss": 0.1569,
+ "step": 766
+ },
+ {
+ "epoch": 19.33,
+ "learning_rate": 0.00019270833333333336,
+ "loss": 0.1628,
+ "step": 767
+ },
+ {
+ "epoch": 19.35,
+ "learning_rate": 0.00019230769230769233,
+ "loss": 0.1529,
+ "step": 768
+ },
+ {
+ "epoch": 19.38,
+ "learning_rate": 0.0001919070512820513,
+ "loss": 0.1606,
+ "step": 769
+ },
+ {
+ "epoch": 19.4,
+ "learning_rate": 0.00019150641025641028,
+ "loss": 0.1564,
+ "step": 770
+ },
+ {
+ "epoch": 19.43,
+ "learning_rate": 0.00019110576923076923,
+ "loss": 0.1599,
+ "step": 771
+ },
+ {
+ "epoch": 19.45,
+ "learning_rate": 0.0001907051282051282,
+ "loss": 0.1653,
+ "step": 772
+ },
+ {
+ "epoch": 19.48,
+ "learning_rate": 0.00019030448717948718,
+ "loss": 0.1587,
+ "step": 773
+ },
+ {
+ "epoch": 19.5,
+ "learning_rate": 0.00018990384615384616,
+ "loss": 0.1683,
+ "step": 774
+ },
+ {
+ "epoch": 19.53,
+ "learning_rate": 0.00018950320512820513,
+ "loss": 0.157,
+ "step": 775
+ },
+ {
+ "epoch": 19.55,
+ "learning_rate": 0.0001891025641025641,
+ "loss": 0.1588,
+ "step": 776
+ },
+ {
+ "epoch": 19.58,
+ "learning_rate": 0.00018870192307692308,
+ "loss": 0.1709,
+ "step": 777
+ },
+ {
+ "epoch": 19.6,
+ "learning_rate": 0.00018830128205128205,
+ "loss": 0.1635,
+ "step": 778
+ },
+ {
+ "epoch": 19.63,
+ "learning_rate": 0.00018790064102564103,
+ "loss": 0.1581,
+ "step": 779
+ },
+ {
+ "epoch": 19.65,
+ "learning_rate": 0.0001875,
+ "loss": 0.1664,
+ "step": 780
+ },
+ {
+ "epoch": 19.68,
+ "learning_rate": 0.00018709935897435898,
+ "loss": 0.1712,
+ "step": 781
+ },
+ {
+ "epoch": 19.7,
+ "learning_rate": 0.00018669871794871795,
+ "loss": 0.165,
+ "step": 782
+ },
+ {
+ "epoch": 19.73,
+ "learning_rate": 0.00018629807692307693,
+ "loss": 0.1692,
+ "step": 783
+ },
+ {
+ "epoch": 19.75,
+ "learning_rate": 0.0001858974358974359,
+ "loss": 0.1658,
+ "step": 784
+ },
+ {
+ "epoch": 19.78,
+ "learning_rate": 0.00018549679487179488,
+ "loss": 0.1702,
+ "step": 785
+ },
+ {
+ "epoch": 19.8,
+ "learning_rate": 0.00018509615384615385,
+ "loss": 0.1672,
+ "step": 786
+ },
+ {
+ "epoch": 19.83,
+ "learning_rate": 0.00018469551282051283,
+ "loss": 0.1708,
+ "step": 787
+ },
+ {
+ "epoch": 19.86,
+ "learning_rate": 0.0001842948717948718,
+ "loss": 0.1705,
+ "step": 788
+ },
+ {
+ "epoch": 19.88,
+ "learning_rate": 0.00018389423076923078,
+ "loss": 0.1853,
+ "step": 789
+ },
+ {
+ "epoch": 19.91,
+ "learning_rate": 0.00018349358974358972,
+ "loss": 0.1701,
+ "step": 790
+ },
+ {
+ "epoch": 19.93,
+ "learning_rate": 0.0001830929487179487,
+ "loss": 0.1726,
+ "step": 791
+ },
+ {
+ "epoch": 19.96,
+ "learning_rate": 0.00018269230769230767,
+ "loss": 0.1829,
+ "step": 792
+ },
+ {
+ "epoch": 19.98,
+ "learning_rate": 0.00018229166666666665,
+ "loss": 0.1749,
+ "step": 793
+ },
+ {
+ "epoch": 20.01,
+ "learning_rate": 0.00018189102564102565,
+ "loss": 0.1559,
+ "step": 794
+ },
+ {
+ "epoch": 20.03,
+ "learning_rate": 0.00018149038461538462,
+ "loss": 0.1424,
+ "step": 795
+ },
+ {
+ "epoch": 20.06,
+ "learning_rate": 0.0001810897435897436,
+ "loss": 0.1386,
+ "step": 796
+ },
+ {
+ "epoch": 20.08,
+ "learning_rate": 0.00018068910256410257,
+ "loss": 0.1388,
+ "step": 797
+ },
+ {
+ "epoch": 20.11,
+ "learning_rate": 0.00018028846153846155,
+ "loss": 0.1385,
+ "step": 798
+ },
+ {
+ "epoch": 20.13,
+ "learning_rate": 0.00017988782051282052,
+ "loss": 0.1349,
+ "step": 799
+ },
+ {
+ "epoch": 20.16,
+ "learning_rate": 0.0001794871794871795,
+ "loss": 0.1357,
+ "step": 800
+ },
+ {
+ "epoch": 20.18,
+ "learning_rate": 0.00017908653846153847,
+ "loss": 0.1384,
+ "step": 801
+ },
+ {
+ "epoch": 20.21,
+ "learning_rate": 0.00017868589743589745,
+ "loss": 0.1425,
+ "step": 802
+ },
+ {
+ "epoch": 20.23,
+ "learning_rate": 0.00017828525641025642,
+ "loss": 0.1337,
+ "step": 803
+ },
+ {
+ "epoch": 20.26,
+ "learning_rate": 0.0001778846153846154,
+ "loss": 0.136,
+ "step": 804
+ },
+ {
+ "epoch": 20.28,
+ "learning_rate": 0.00017748397435897437,
+ "loss": 0.1382,
+ "step": 805
+ },
+ {
+ "epoch": 20.31,
+ "learning_rate": 0.00017708333333333335,
+ "loss": 0.1396,
+ "step": 806
+ },
+ {
+ "epoch": 20.33,
+ "learning_rate": 0.00017668269230769232,
+ "loss": 0.1411,
+ "step": 807
+ },
+ {
+ "epoch": 20.36,
+ "learning_rate": 0.0001762820512820513,
+ "loss": 0.1375,
+ "step": 808
+ },
+ {
+ "epoch": 20.38,
+ "learning_rate": 0.00017588141025641027,
+ "loss": 0.1457,
+ "step": 809
+ },
+ {
+ "epoch": 20.41,
+ "learning_rate": 0.00017548076923076922,
+ "loss": 0.1447,
+ "step": 810
+ },
+ {
+ "epoch": 20.43,
+ "learning_rate": 0.0001750801282051282,
+ "loss": 0.1381,
+ "step": 811
+ },
+ {
+ "epoch": 20.46,
+ "learning_rate": 0.00017467948717948717,
+ "loss": 0.1443,
+ "step": 812
+ },
+ {
+ "epoch": 20.49,
+ "learning_rate": 0.00017427884615384614,
+ "loss": 0.1427,
+ "step": 813
+ },
+ {
+ "epoch": 20.51,
+ "learning_rate": 0.00017387820512820512,
+ "loss": 0.1425,
+ "step": 814
+ },
+ {
+ "epoch": 20.54,
+ "learning_rate": 0.0001734775641025641,
+ "loss": 0.1428,
+ "step": 815
+ },
+ {
+ "epoch": 20.56,
+ "learning_rate": 0.00017307692307692307,
+ "loss": 0.1454,
+ "step": 816
+ },
+ {
+ "epoch": 20.59,
+ "learning_rate": 0.00017267628205128204,
+ "loss": 0.1422,
+ "step": 817
+ },
+ {
+ "epoch": 20.61,
+ "learning_rate": 0.00017227564102564102,
+ "loss": 0.1426,
+ "step": 818
+ },
+ {
+ "epoch": 20.64,
+ "learning_rate": 0.000171875,
+ "loss": 0.1453,
+ "step": 819
+ },
+ {
+ "epoch": 20.66,
+ "learning_rate": 0.000171474358974359,
+ "loss": 0.1472,
+ "step": 820
+ },
+ {
+ "epoch": 20.69,
+ "learning_rate": 0.00017107371794871797,
+ "loss": 0.1479,
+ "step": 821
+ },
+ {
+ "epoch": 20.71,
+ "learning_rate": 0.00017067307692307694,
+ "loss": 0.1387,
+ "step": 822
+ },
+ {
+ "epoch": 20.74,
+ "learning_rate": 0.00017027243589743592,
+ "loss": 0.15,
+ "step": 823
+ },
+ {
+ "epoch": 20.76,
+ "learning_rate": 0.0001698717948717949,
+ "loss": 0.1429,
+ "step": 824
+ },
+ {
+ "epoch": 20.79,
+ "learning_rate": 0.00016947115384615387,
+ "loss": 0.1532,
+ "step": 825
+ },
+ {
+ "epoch": 20.81,
+ "learning_rate": 0.00016907051282051284,
+ "loss": 0.1452,
+ "step": 826
+ },
+ {
+ "epoch": 20.84,
+ "learning_rate": 0.00016866987179487181,
+ "loss": 0.1445,
+ "step": 827
+ },
+ {
+ "epoch": 20.86,
+ "learning_rate": 0.0001682692307692308,
+ "loss": 0.1461,
+ "step": 828
+ },
+ {
+ "epoch": 20.89,
+ "learning_rate": 0.00016786858974358974,
+ "loss": 0.1544,
+ "step": 829
+ },
+ {
+ "epoch": 20.91,
+ "learning_rate": 0.0001674679487179487,
+ "loss": 0.1476,
+ "step": 830
+ },
+ {
+ "epoch": 20.94,
+ "learning_rate": 0.00016706730769230769,
+ "loss": 0.1535,
+ "step": 831
+ },
+ {
+ "epoch": 20.96,
+ "learning_rate": 0.00016666666666666666,
+ "loss": 0.1498,
+ "step": 832
+ },
+ {
+ "epoch": 20.99,
+ "learning_rate": 0.00016626602564102564,
+ "loss": 0.1514,
+ "step": 833
+ },
+ {
+ "epoch": 21.01,
+ "learning_rate": 0.0001658653846153846,
+ "loss": 0.1405,
+ "step": 834
+ },
+ {
+ "epoch": 21.04,
+ "learning_rate": 0.00016546474358974359,
+ "loss": 0.1218,
+ "step": 835
+ },
+ {
+ "epoch": 21.06,
+ "learning_rate": 0.00016506410256410256,
+ "loss": 0.1228,
+ "step": 836
+ },
+ {
+ "epoch": 21.09,
+ "learning_rate": 0.00016466346153846153,
+ "loss": 0.119,
+ "step": 837
+ },
+ {
+ "epoch": 21.11,
+ "learning_rate": 0.0001642628205128205,
+ "loss": 0.1278,
+ "step": 838
+ },
+ {
+ "epoch": 21.14,
+ "learning_rate": 0.00016386217948717948,
+ "loss": 0.1205,
+ "step": 839
+ },
+ {
+ "epoch": 21.17,
+ "learning_rate": 0.00016346153846153846,
+ "loss": 0.1234,
+ "step": 840
+ },
+ {
+ "epoch": 21.19,
+ "learning_rate": 0.00016306089743589743,
+ "loss": 0.1226,
+ "step": 841
+ },
+ {
+ "epoch": 21.22,
+ "learning_rate": 0.0001626602564102564,
+ "loss": 0.1216,
+ "step": 842
+ },
+ {
+ "epoch": 21.24,
+ "learning_rate": 0.00016225961538461538,
+ "loss": 0.1221,
+ "step": 843
+ },
+ {
+ "epoch": 21.27,
+ "learning_rate": 0.00016185897435897436,
+ "loss": 0.1234,
+ "step": 844
+ },
+ {
+ "epoch": 21.29,
+ "learning_rate": 0.00016145833333333336,
+ "loss": 0.1226,
+ "step": 845
+ },
+ {
+ "epoch": 21.32,
+ "learning_rate": 0.00016105769230769233,
+ "loss": 0.1224,
+ "step": 846
+ },
+ {
+ "epoch": 21.34,
+ "learning_rate": 0.0001606570512820513,
+ "loss": 0.1266,
+ "step": 847
+ },
+ {
+ "epoch": 21.37,
+ "learning_rate": 0.00016025641025641028,
+ "loss": 0.1267,
+ "step": 848
+ },
+ {
+ "epoch": 21.39,
+ "learning_rate": 0.00015985576923076923,
+ "loss": 0.129,
+ "step": 849
+ },
+ {
+ "epoch": 21.42,
+ "learning_rate": 0.0001594551282051282,
+ "loss": 0.1213,
+ "step": 850
+ },
+ {
+ "epoch": 21.44,
+ "learning_rate": 0.00015905448717948718,
+ "loss": 0.1273,
+ "step": 851
+ },
+ {
+ "epoch": 21.47,
+ "learning_rate": 0.00015865384615384616,
+ "loss": 0.1352,
+ "step": 852
+ },
+ {
+ "epoch": 21.49,
+ "learning_rate": 0.00015825320512820513,
+ "loss": 0.1328,
+ "step": 853
+ },
+ {
+ "epoch": 21.52,
+ "learning_rate": 0.0001578525641025641,
+ "loss": 0.1253,
+ "step": 854
+ },
+ {
+ "epoch": 21.54,
+ "learning_rate": 0.00015745192307692308,
+ "loss": 0.1267,
+ "step": 855
+ },
+ {
+ "epoch": 21.57,
+ "learning_rate": 0.00015705128205128205,
+ "loss": 0.1325,
+ "step": 856
+ },
+ {
+ "epoch": 21.59,
+ "learning_rate": 0.00015665064102564103,
+ "loss": 0.1267,
+ "step": 857
+ },
+ {
+ "epoch": 21.62,
+ "learning_rate": 0.00015625,
+ "loss": 0.1326,
+ "step": 858
+ },
+ {
+ "epoch": 21.64,
+ "learning_rate": 0.00015584935897435898,
+ "loss": 0.1319,
+ "step": 859
+ },
+ {
+ "epoch": 21.67,
+ "learning_rate": 0.00015544871794871795,
+ "loss": 0.1262,
+ "step": 860
+ },
+ {
+ "epoch": 21.69,
+ "learning_rate": 0.00015504807692307693,
+ "loss": 0.1334,
+ "step": 861
+ },
+ {
+ "epoch": 21.72,
+ "learning_rate": 0.0001546474358974359,
+ "loss": 0.1347,
+ "step": 862
+ },
+ {
+ "epoch": 21.74,
+ "learning_rate": 0.00015424679487179488,
+ "loss": 0.1322,
+ "step": 863
+ },
+ {
+ "epoch": 21.77,
+ "learning_rate": 0.00015384615384615385,
+ "loss": 0.1321,
+ "step": 864
+ },
+ {
+ "epoch": 21.8,
+ "learning_rate": 0.00015344551282051283,
+ "loss": 0.128,
+ "step": 865
+ },
+ {
+ "epoch": 21.82,
+ "learning_rate": 0.0001530448717948718,
+ "loss": 0.1365,
+ "step": 866
+ },
+ {
+ "epoch": 21.85,
+ "learning_rate": 0.00015264423076923078,
+ "loss": 0.137,
+ "step": 867
+ },
+ {
+ "epoch": 21.87,
+ "learning_rate": 0.00015224358974358972,
+ "loss": 0.1322,
+ "step": 868
+ },
+ {
+ "epoch": 21.9,
+ "learning_rate": 0.0001518429487179487,
+ "loss": 0.1341,
+ "step": 869
+ },
+ {
+ "epoch": 21.92,
+ "learning_rate": 0.00015144230769230767,
+ "loss": 0.1334,
+ "step": 870
+ },
+ {
+ "epoch": 21.95,
+ "learning_rate": 0.00015104166666666665,
+ "loss": 0.1339,
+ "step": 871
+ },
+ {
+ "epoch": 21.97,
+ "learning_rate": 0.00015064102564102565,
+ "loss": 0.1301,
+ "step": 872
+ },
+ {
+ "epoch": 22.0,
+ "learning_rate": 0.00015024038461538462,
+ "loss": 0.1325,
+ "step": 873
+ },
+ {
+ "epoch": 22.02,
+ "learning_rate": 0.0001498397435897436,
+ "loss": 0.1153,
+ "step": 874
+ },
+ {
+ "epoch": 22.05,
+ "learning_rate": 0.00014943910256410257,
+ "loss": 0.1113,
+ "step": 875
+ },
+ {
+ "epoch": 22.07,
+ "learning_rate": 0.00014903846153846155,
+ "loss": 0.1138,
+ "step": 876
+ },
+ {
+ "epoch": 22.1,
+ "learning_rate": 0.00014863782051282052,
+ "loss": 0.1136,
+ "step": 877
+ },
+ {
+ "epoch": 22.12,
+ "learning_rate": 0.0001482371794871795,
+ "loss": 0.1096,
+ "step": 878
+ },
+ {
+ "epoch": 22.15,
+ "learning_rate": 0.00014783653846153847,
+ "loss": 0.116,
+ "step": 879
+ },
+ {
+ "epoch": 22.17,
+ "learning_rate": 0.00014743589743589745,
+ "loss": 0.1122,
+ "step": 880
+ },
+ {
+ "epoch": 22.2,
+ "learning_rate": 0.00014703525641025642,
+ "loss": 0.1132,
+ "step": 881
+ },
+ {
+ "epoch": 22.22,
+ "learning_rate": 0.0001466346153846154,
+ "loss": 0.1128,
+ "step": 882
+ },
+ {
+ "epoch": 22.25,
+ "learning_rate": 0.00014623397435897437,
+ "loss": 0.1134,
+ "step": 883
+ },
+ {
+ "epoch": 22.27,
+ "learning_rate": 0.00014583333333333335,
+ "loss": 0.1065,
+ "step": 884
+ },
+ {
+ "epoch": 22.3,
+ "learning_rate": 0.00014543269230769232,
+ "loss": 0.117,
+ "step": 885
+ },
+ {
+ "epoch": 22.32,
+ "learning_rate": 0.0001450320512820513,
+ "loss": 0.112,
+ "step": 886
+ },
+ {
+ "epoch": 22.35,
+ "learning_rate": 0.00014463141025641027,
+ "loss": 0.1145,
+ "step": 887
+ },
+ {
+ "epoch": 22.37,
+ "learning_rate": 0.00014423076923076922,
+ "loss": 0.1128,
+ "step": 888
+ },
+ {
+ "epoch": 22.4,
+ "learning_rate": 0.0001438301282051282,
+ "loss": 0.1132,
+ "step": 889
+ },
+ {
+ "epoch": 22.43,
+ "learning_rate": 0.00014342948717948717,
+ "loss": 0.1143,
+ "step": 890
+ },
+ {
+ "epoch": 22.45,
+ "learning_rate": 0.00014302884615384614,
+ "loss": 0.1172,
+ "step": 891
+ },
+ {
+ "epoch": 22.48,
+ "learning_rate": 0.00014262820512820512,
+ "loss": 0.1143,
+ "step": 892
+ },
+ {
+ "epoch": 22.5,
+ "learning_rate": 0.0001422275641025641,
+ "loss": 0.1164,
+ "step": 893
+ },
+ {
+ "epoch": 22.53,
+ "learning_rate": 0.00014182692307692307,
+ "loss": 0.1197,
+ "step": 894
+ },
+ {
+ "epoch": 22.55,
+ "learning_rate": 0.00014142628205128204,
+ "loss": 0.1152,
+ "step": 895
+ },
+ {
+ "epoch": 22.58,
+ "learning_rate": 0.00014102564102564101,
+ "loss": 0.1243,
+ "step": 896
+ },
+ {
+ "epoch": 22.6,
+ "learning_rate": 0.00014062500000000002,
+ "loss": 0.1167,
+ "step": 897
+ },
+ {
+ "epoch": 22.63,
+ "learning_rate": 0.000140224358974359,
+ "loss": 0.1224,
+ "step": 898
+ },
+ {
+ "epoch": 22.65,
+ "learning_rate": 0.00013982371794871797,
+ "loss": 0.1192,
+ "step": 899
+ },
+ {
+ "epoch": 22.68,
+ "learning_rate": 0.00013942307692307694,
+ "loss": 0.1184,
+ "step": 900
+ },
+ {
+ "epoch": 22.7,
+ "learning_rate": 0.00013902243589743592,
+ "loss": 0.1169,
+ "step": 901
+ },
+ {
+ "epoch": 22.73,
+ "learning_rate": 0.0001386217948717949,
+ "loss": 0.1169,
+ "step": 902
+ },
+ {
+ "epoch": 22.75,
+ "learning_rate": 0.00013822115384615386,
+ "loss": 0.1177,
+ "step": 903
+ },
+ {
+ "epoch": 22.78,
+ "learning_rate": 0.00013782051282051284,
+ "loss": 0.1175,
+ "step": 904
+ },
+ {
+ "epoch": 22.8,
+ "learning_rate": 0.00013741987179487181,
+ "loss": 0.1164,
+ "step": 905
+ },
+ {
+ "epoch": 22.83,
+ "learning_rate": 0.0001370192307692308,
+ "loss": 0.1215,
+ "step": 906
+ },
+ {
+ "epoch": 22.85,
+ "learning_rate": 0.00013661858974358974,
+ "loss": 0.1204,
+ "step": 907
+ },
+ {
+ "epoch": 22.88,
+ "learning_rate": 0.0001362179487179487,
+ "loss": 0.1208,
+ "step": 908
+ },
+ {
+ "epoch": 22.9,
+ "learning_rate": 0.00013581730769230769,
+ "loss": 0.1175,
+ "step": 909
+ },
+ {
+ "epoch": 22.93,
+ "learning_rate": 0.00013541666666666666,
+ "loss": 0.1214,
+ "step": 910
+ },
+ {
+ "epoch": 22.95,
+ "learning_rate": 0.00013501602564102564,
+ "loss": 0.1243,
+ "step": 911
+ },
+ {
+ "epoch": 22.98,
+ "learning_rate": 0.0001346153846153846,
+ "loss": 0.1266,
+ "step": 912
+ },
+ {
+ "epoch": 23.0,
+ "learning_rate": 0.00013421474358974358,
+ "loss": 0.1226,
+ "step": 913
+ },
+ {
+ "epoch": 23.03,
+ "learning_rate": 0.00013381410256410256,
+ "loss": 0.1058,
+ "step": 914
+ },
+ {
+ "epoch": 23.06,
+ "learning_rate": 0.00013341346153846153,
+ "loss": 0.1034,
+ "step": 915
+ },
+ {
+ "epoch": 23.08,
+ "learning_rate": 0.0001330128205128205,
+ "loss": 0.1026,
+ "step": 916
+ },
+ {
+ "epoch": 23.11,
+ "learning_rate": 0.00013261217948717948,
+ "loss": 0.1026,
+ "step": 917
+ },
+ {
+ "epoch": 23.13,
+ "learning_rate": 0.00013221153846153846,
+ "loss": 0.1028,
+ "step": 918
+ },
+ {
+ "epoch": 23.16,
+ "learning_rate": 0.00013181089743589743,
+ "loss": 0.1024,
+ "step": 919
+ },
+ {
+ "epoch": 23.18,
+ "learning_rate": 0.0001314102564102564,
+ "loss": 0.1042,
+ "step": 920
+ },
+ {
+ "epoch": 23.21,
+ "learning_rate": 0.00013100961538461538,
+ "loss": 0.0992,
+ "step": 921
+ },
+ {
+ "epoch": 23.23,
+ "learning_rate": 0.00013060897435897436,
+ "loss": 0.1048,
+ "step": 922
+ },
+ {
+ "epoch": 23.26,
+ "learning_rate": 0.00013020833333333336,
+ "loss": 0.1006,
+ "step": 923
+ },
+ {
+ "epoch": 23.28,
+ "learning_rate": 0.00012980769230769233,
+ "loss": 0.1041,
+ "step": 924
+ },
+ {
+ "epoch": 23.31,
+ "learning_rate": 0.0001294070512820513,
+ "loss": 0.1036,
+ "step": 925
+ },
+ {
+ "epoch": 23.33,
+ "learning_rate": 0.00012900641025641028,
+ "loss": 0.1045,
+ "step": 926
+ },
+ {
+ "epoch": 23.36,
+ "learning_rate": 0.00012860576923076923,
+ "loss": 0.1048,
+ "step": 927
+ },
+ {
+ "epoch": 23.38,
+ "learning_rate": 0.0001282051282051282,
+ "loss": 0.1082,
+ "step": 928
+ },
+ {
+ "epoch": 23.41,
+ "learning_rate": 0.00012780448717948718,
+ "loss": 0.1092,
+ "step": 929
+ },
+ {
+ "epoch": 23.43,
+ "learning_rate": 0.00012740384615384615,
+ "loss": 0.1085,
+ "step": 930
+ },
+ {
+ "epoch": 23.46,
+ "learning_rate": 0.00012700320512820513,
+ "loss": 0.1037,
+ "step": 931
+ },
+ {
+ "epoch": 23.48,
+ "learning_rate": 0.0001266025641025641,
+ "loss": 0.109,
+ "step": 932
+ },
+ {
+ "epoch": 23.51,
+ "learning_rate": 0.00012620192307692308,
+ "loss": 0.1076,
+ "step": 933
+ },
+ {
+ "epoch": 23.53,
+ "learning_rate": 0.00012580128205128205,
+ "loss": 0.1062,
+ "step": 934
+ },
+ {
+ "epoch": 23.56,
+ "learning_rate": 0.00012540064102564103,
+ "loss": 0.1049,
+ "step": 935
+ },
+ {
+ "epoch": 23.58,
+ "learning_rate": 0.000125,
+ "loss": 0.1057,
+ "step": 936
+ },
+ {
+ "epoch": 23.61,
+ "learning_rate": 0.00012459935897435898,
+ "loss": 0.1106,
+ "step": 937
+ },
+ {
+ "epoch": 23.63,
+ "learning_rate": 0.00012419871794871795,
+ "loss": 0.1134,
+ "step": 938
+ },
+ {
+ "epoch": 23.66,
+ "learning_rate": 0.00012379807692307693,
+ "loss": 0.1124,
+ "step": 939
+ },
+ {
+ "epoch": 23.69,
+ "learning_rate": 0.0001233974358974359,
+ "loss": 0.105,
+ "step": 940
+ },
+ {
+ "epoch": 23.71,
+ "learning_rate": 0.00012299679487179488,
+ "loss": 0.113,
+ "step": 941
+ },
+ {
+ "epoch": 23.74,
+ "learning_rate": 0.00012259615384615385,
+ "loss": 0.1104,
+ "step": 942
+ },
+ {
+ "epoch": 23.76,
+ "learning_rate": 0.00012219551282051283,
+ "loss": 0.1136,
+ "step": 943
+ },
+ {
+ "epoch": 23.79,
+ "learning_rate": 0.0001217948717948718,
+ "loss": 0.1104,
+ "step": 944
+ },
+ {
+ "epoch": 23.81,
+ "learning_rate": 0.00012139423076923077,
+ "loss": 0.1054,
+ "step": 945
+ },
+ {
+ "epoch": 23.84,
+ "learning_rate": 0.00012099358974358975,
+ "loss": 0.1109,
+ "step": 946
+ },
+ {
+ "epoch": 23.86,
+ "learning_rate": 0.00012059294871794872,
+ "loss": 0.1062,
+ "step": 947
+ },
+ {
+ "epoch": 23.89,
+ "learning_rate": 0.0001201923076923077,
+ "loss": 0.1185,
+ "step": 948
+ },
+ {
+ "epoch": 23.91,
+ "learning_rate": 0.00011979166666666667,
+ "loss": 0.1138,
+ "step": 949
+ },
+ {
+ "epoch": 23.94,
+ "learning_rate": 0.00011939102564102565,
+ "loss": 0.1119,
+ "step": 950
+ },
+ {
+ "epoch": 23.96,
+ "learning_rate": 0.00011899038461538461,
+ "loss": 0.1114,
+ "step": 951
+ },
+ {
+ "epoch": 23.99,
+ "learning_rate": 0.00011858974358974358,
+ "loss": 0.1122,
+ "step": 952
+ },
+ {
+ "epoch": 24.01,
+ "learning_rate": 0.00011818910256410256,
+ "loss": 0.1051,
+ "step": 953
+ },
+ {
+ "epoch": 24.04,
+ "learning_rate": 0.00011778846153846153,
+ "loss": 0.0924,
+ "step": 954
+ },
+ {
+ "epoch": 24.06,
+ "learning_rate": 0.00011738782051282051,
+ "loss": 0.0977,
+ "step": 955
+ },
+ {
+ "epoch": 24.09,
+ "learning_rate": 0.0001169871794871795,
+ "loss": 0.0945,
+ "step": 956
+ },
+ {
+ "epoch": 24.11,
+ "learning_rate": 0.00011658653846153847,
+ "loss": 0.0958,
+ "step": 957
+ },
+ {
+ "epoch": 24.14,
+ "learning_rate": 0.00011618589743589745,
+ "loss": 0.0934,
+ "step": 958
+ },
+ {
+ "epoch": 24.16,
+ "learning_rate": 0.00011578525641025642,
+ "loss": 0.0952,
+ "step": 959
+ },
+ {
+ "epoch": 24.19,
+ "learning_rate": 0.0001153846153846154,
+ "loss": 0.0993,
+ "step": 960
+ },
+ {
+ "epoch": 24.21,
+ "learning_rate": 0.00011498397435897436,
+ "loss": 0.0978,
+ "step": 961
+ },
+ {
+ "epoch": 24.24,
+ "learning_rate": 0.00011458333333333333,
+ "loss": 0.0984,
+ "step": 962
+ },
+ {
+ "epoch": 24.26,
+ "learning_rate": 0.0001141826923076923,
+ "loss": 0.1019,
+ "step": 963
+ },
+ {
+ "epoch": 24.29,
+ "learning_rate": 0.00011378205128205128,
+ "loss": 0.0994,
+ "step": 964
+ },
+ {
+ "epoch": 24.31,
+ "learning_rate": 0.00011338141025641026,
+ "loss": 0.1006,
+ "step": 965
+ },
+ {
+ "epoch": 24.34,
+ "learning_rate": 0.00011298076923076923,
+ "loss": 0.0971,
+ "step": 966
+ },
+ {
+ "epoch": 24.37,
+ "learning_rate": 0.0001125801282051282,
+ "loss": 0.1008,
+ "step": 967
+ },
+ {
+ "epoch": 24.39,
+ "learning_rate": 0.00011217948717948718,
+ "loss": 0.0948,
+ "step": 968
+ },
+ {
+ "epoch": 24.42,
+ "learning_rate": 0.00011177884615384617,
+ "loss": 0.1026,
+ "step": 969
+ },
+ {
+ "epoch": 24.44,
+ "learning_rate": 0.00011137820512820514,
+ "loss": 0.0966,
+ "step": 970
+ },
+ {
+ "epoch": 24.47,
+ "learning_rate": 0.0001109775641025641,
+ "loss": 0.1004,
+ "step": 971
+ },
+ {
+ "epoch": 24.49,
+ "learning_rate": 0.00011057692307692308,
+ "loss": 0.1031,
+ "step": 972
+ },
+ {
+ "epoch": 24.52,
+ "learning_rate": 0.00011017628205128205,
+ "loss": 0.0984,
+ "step": 973
+ },
+ {
+ "epoch": 24.54,
+ "learning_rate": 0.00010977564102564103,
+ "loss": 0.0988,
+ "step": 974
+ },
+ {
+ "epoch": 24.57,
+ "learning_rate": 0.000109375,
+ "loss": 0.1048,
+ "step": 975
+ },
+ {
+ "epoch": 24.59,
+ "learning_rate": 0.00010897435897435898,
+ "loss": 0.0987,
+ "step": 976
+ },
+ {
+ "epoch": 24.62,
+ "learning_rate": 0.00010857371794871795,
+ "loss": 0.1026,
+ "step": 977
+ },
+ {
+ "epoch": 24.64,
+ "learning_rate": 0.00010817307692307693,
+ "loss": 0.1015,
+ "step": 978
+ },
+ {
+ "epoch": 24.67,
+ "learning_rate": 0.0001077724358974359,
+ "loss": 0.102,
+ "step": 979
+ },
+ {
+ "epoch": 24.69,
+ "learning_rate": 0.00010737179487179486,
+ "loss": 0.1005,
+ "step": 980
+ },
+ {
+ "epoch": 24.72,
+ "learning_rate": 0.00010697115384615384,
+ "loss": 0.1042,
+ "step": 981
+ },
+ {
+ "epoch": 24.74,
+ "learning_rate": 0.00010657051282051283,
+ "loss": 0.1046,
+ "step": 982
+ },
+ {
+ "epoch": 24.77,
+ "learning_rate": 0.0001061698717948718,
+ "loss": 0.0976,
+ "step": 983
+ },
+ {
+ "epoch": 24.79,
+ "learning_rate": 0.00010576923076923077,
+ "loss": 0.1014,
+ "step": 984
+ },
+ {
+ "epoch": 24.82,
+ "learning_rate": 0.00010536858974358975,
+ "loss": 0.1028,
+ "step": 985
+ },
+ {
+ "epoch": 24.84,
+ "learning_rate": 0.00010496794871794872,
+ "loss": 0.1022,
+ "step": 986
+ },
+ {
+ "epoch": 24.87,
+ "learning_rate": 0.0001045673076923077,
+ "loss": 0.1021,
+ "step": 987
+ },
+ {
+ "epoch": 24.89,
+ "learning_rate": 0.00010416666666666667,
+ "loss": 0.1047,
+ "step": 988
+ },
+ {
+ "epoch": 24.92,
+ "learning_rate": 0.00010376602564102565,
+ "loss": 0.1062,
+ "step": 989
+ },
+ {
+ "epoch": 24.94,
+ "learning_rate": 0.00010336538461538461,
+ "loss": 0.1028,
+ "step": 990
+ },
+ {
+ "epoch": 24.97,
+ "learning_rate": 0.00010296474358974358,
+ "loss": 0.1051,
+ "step": 991
+ },
+ {
+ "epoch": 25.0,
+ "learning_rate": 0.00010256410256410256,
+ "loss": 0.1047,
+ "step": 992
+ },
+ {
+ "epoch": 25.02,
+ "learning_rate": 0.00010216346153846153,
+ "loss": 0.097,
+ "step": 993
+ },
+ {
+ "epoch": 25.05,
+ "learning_rate": 0.00010176282051282051,
+ "loss": 0.0963,
+ "step": 994
+ },
+ {
+ "epoch": 25.07,
+ "learning_rate": 0.0001013621794871795,
+ "loss": 0.0888,
+ "step": 995
+ },
+ {
+ "epoch": 25.1,
+ "learning_rate": 0.00010096153846153847,
+ "loss": 0.089,
+ "step": 996
+ },
+ {
+ "epoch": 25.12,
+ "learning_rate": 0.00010056089743589745,
+ "loss": 0.0911,
+ "step": 997
+ },
+ {
+ "epoch": 25.15,
+ "learning_rate": 0.00010016025641025642,
+ "loss": 0.0916,
+ "step": 998
+ },
+ {
+ "epoch": 25.17,
+ "learning_rate": 9.97596153846154e-05,
+ "loss": 0.089,
+ "step": 999
+ },
+ {
+ "epoch": 25.2,
+ "learning_rate": 9.935897435897436e-05,
+ "loss": 0.0883,
+ "step": 1000
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 5.630264794033152e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-1000/training_args.bin b/checkpoint-1000/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-1000/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-1100/README.md b/checkpoint-1100/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-1100/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-1100/adapter_config.json b/checkpoint-1100/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-1100/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-1100/adapter_model.bin b/checkpoint-1100/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..9d031b7584b3be6d34d9f461ef580ee78bfb45f4
--- /dev/null
+++ b/checkpoint-1100/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6c88ad543ad0cd581cce028fd3dc8d689a3d5d1f1655d648f44d44b3f227ec50
+size 39409357
diff --git a/checkpoint-1100/optimizer.pt b/checkpoint-1100/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..53d4ba36c4a3b50b9ceba7d12f49bce2e90abb99
--- /dev/null
+++ b/checkpoint-1100/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5b80027248fc1a38c437e756d546038dbc041930932c6a4ce5ee2f0941b779e7
+size 78844421
diff --git a/checkpoint-1100/rng_state.pth b/checkpoint-1100/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8a1d2b6ce45c272f08235cf96219814508f02912
--- /dev/null
+++ b/checkpoint-1100/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:82632deb3eef79ffa0456776bb9d6e5a9a57b9d7022aadd020f4d487704bacab
+size 14575
diff --git a/checkpoint-1100/scheduler.pt b/checkpoint-1100/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..a563aeb7e8a9556376fa7b81f457bc44d1241bf1
--- /dev/null
+++ b/checkpoint-1100/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8dc44d9525e28449178817fd7456c24041b688fdd22e4d5015874b3026a0e057
+size 627
diff --git a/checkpoint-1100/special_tokens_map.json b/checkpoint-1100/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-1100/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-1100/tokenizer.json b/checkpoint-1100/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-1100/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-1100/tokenizer_config.json b/checkpoint-1100/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-1100/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-1100/trainer_state.json b/checkpoint-1100/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..3c15198d5788dbd7e9ce509e50a3d21de62392f0
--- /dev/null
+++ b/checkpoint-1100/trainer_state.json
@@ -0,0 +1,6619 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 27.716535433070867,
+ "eval_steps": 500,
+ "global_step": 1100,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ },
+ {
+ "epoch": 10.1,
+ "learning_rate": 0.00033934294871794873,
+ "loss": 0.7635,
+ "step": 401
+ },
+ {
+ "epoch": 10.13,
+ "learning_rate": 0.00033894230769230773,
+ "loss": 0.7713,
+ "step": 402
+ },
+ {
+ "epoch": 10.15,
+ "learning_rate": 0.0003385416666666667,
+ "loss": 0.827,
+ "step": 403
+ },
+ {
+ "epoch": 10.18,
+ "learning_rate": 0.0003381410256410257,
+ "loss": 0.7844,
+ "step": 404
+ },
+ {
+ "epoch": 10.2,
+ "learning_rate": 0.00033774038461538463,
+ "loss": 0.7839,
+ "step": 405
+ },
+ {
+ "epoch": 10.23,
+ "learning_rate": 0.00033733974358974363,
+ "loss": 0.7702,
+ "step": 406
+ },
+ {
+ "epoch": 10.26,
+ "learning_rate": 0.0003369391025641026,
+ "loss": 0.7855,
+ "step": 407
+ },
+ {
+ "epoch": 10.28,
+ "learning_rate": 0.0003365384615384616,
+ "loss": 0.7669,
+ "step": 408
+ },
+ {
+ "epoch": 10.31,
+ "learning_rate": 0.0003361378205128205,
+ "loss": 0.8054,
+ "step": 409
+ },
+ {
+ "epoch": 10.33,
+ "learning_rate": 0.0003357371794871795,
+ "loss": 0.7792,
+ "step": 410
+ },
+ {
+ "epoch": 10.36,
+ "learning_rate": 0.0003353365384615384,
+ "loss": 0.7986,
+ "step": 411
+ },
+ {
+ "epoch": 10.38,
+ "learning_rate": 0.0003349358974358974,
+ "loss": 0.8269,
+ "step": 412
+ },
+ {
+ "epoch": 10.41,
+ "learning_rate": 0.00033453525641025637,
+ "loss": 0.7923,
+ "step": 413
+ },
+ {
+ "epoch": 10.43,
+ "learning_rate": 0.00033413461538461537,
+ "loss": 0.8242,
+ "step": 414
+ },
+ {
+ "epoch": 10.46,
+ "learning_rate": 0.0003337339743589743,
+ "loss": 0.8496,
+ "step": 415
+ },
+ {
+ "epoch": 10.48,
+ "learning_rate": 0.0003333333333333333,
+ "loss": 0.8236,
+ "step": 416
+ },
+ {
+ "epoch": 10.51,
+ "learning_rate": 0.0003329326923076923,
+ "loss": 0.8685,
+ "step": 417
+ },
+ {
+ "epoch": 10.53,
+ "learning_rate": 0.00033253205128205127,
+ "loss": 0.8542,
+ "step": 418
+ },
+ {
+ "epoch": 10.56,
+ "learning_rate": 0.0003321314102564103,
+ "loss": 0.8185,
+ "step": 419
+ },
+ {
+ "epoch": 10.58,
+ "learning_rate": 0.0003317307692307692,
+ "loss": 0.8477,
+ "step": 420
+ },
+ {
+ "epoch": 10.61,
+ "learning_rate": 0.0003313301282051282,
+ "loss": 0.8372,
+ "step": 421
+ },
+ {
+ "epoch": 10.63,
+ "learning_rate": 0.00033092948717948717,
+ "loss": 0.8731,
+ "step": 422
+ },
+ {
+ "epoch": 10.66,
+ "learning_rate": 0.00033052884615384617,
+ "loss": 0.8322,
+ "step": 423
+ },
+ {
+ "epoch": 10.68,
+ "learning_rate": 0.0003301282051282051,
+ "loss": 0.8514,
+ "step": 424
+ },
+ {
+ "epoch": 10.71,
+ "learning_rate": 0.0003297275641025641,
+ "loss": 0.8564,
+ "step": 425
+ },
+ {
+ "epoch": 10.73,
+ "learning_rate": 0.00032932692307692307,
+ "loss": 0.8278,
+ "step": 426
+ },
+ {
+ "epoch": 10.76,
+ "learning_rate": 0.00032892628205128207,
+ "loss": 0.8408,
+ "step": 427
+ },
+ {
+ "epoch": 10.78,
+ "learning_rate": 0.000328525641025641,
+ "loss": 0.8438,
+ "step": 428
+ },
+ {
+ "epoch": 10.81,
+ "learning_rate": 0.000328125,
+ "loss": 0.8862,
+ "step": 429
+ },
+ {
+ "epoch": 10.83,
+ "learning_rate": 0.00032772435897435897,
+ "loss": 0.8631,
+ "step": 430
+ },
+ {
+ "epoch": 10.86,
+ "learning_rate": 0.00032732371794871797,
+ "loss": 0.883,
+ "step": 431
+ },
+ {
+ "epoch": 10.89,
+ "learning_rate": 0.0003269230769230769,
+ "loss": 0.8644,
+ "step": 432
+ },
+ {
+ "epoch": 10.91,
+ "learning_rate": 0.0003265224358974359,
+ "loss": 0.8907,
+ "step": 433
+ },
+ {
+ "epoch": 10.94,
+ "learning_rate": 0.00032612179487179487,
+ "loss": 0.9141,
+ "step": 434
+ },
+ {
+ "epoch": 10.96,
+ "learning_rate": 0.00032572115384615387,
+ "loss": 0.8877,
+ "step": 435
+ },
+ {
+ "epoch": 10.99,
+ "learning_rate": 0.0003253205128205128,
+ "loss": 0.8521,
+ "step": 436
+ },
+ {
+ "epoch": 11.01,
+ "learning_rate": 0.0003249198717948718,
+ "loss": 0.7842,
+ "step": 437
+ },
+ {
+ "epoch": 11.04,
+ "learning_rate": 0.00032451923076923077,
+ "loss": 0.6516,
+ "step": 438
+ },
+ {
+ "epoch": 11.06,
+ "learning_rate": 0.00032411858974358977,
+ "loss": 0.6549,
+ "step": 439
+ },
+ {
+ "epoch": 11.09,
+ "learning_rate": 0.0003237179487179487,
+ "loss": 0.6736,
+ "step": 440
+ },
+ {
+ "epoch": 11.11,
+ "learning_rate": 0.0003233173076923077,
+ "loss": 0.6625,
+ "step": 441
+ },
+ {
+ "epoch": 11.14,
+ "learning_rate": 0.0003229166666666667,
+ "loss": 0.6466,
+ "step": 442
+ },
+ {
+ "epoch": 11.16,
+ "learning_rate": 0.00032251602564102567,
+ "loss": 0.6362,
+ "step": 443
+ },
+ {
+ "epoch": 11.19,
+ "learning_rate": 0.00032211538461538467,
+ "loss": 0.674,
+ "step": 444
+ },
+ {
+ "epoch": 11.21,
+ "learning_rate": 0.0003217147435897436,
+ "loss": 0.6509,
+ "step": 445
+ },
+ {
+ "epoch": 11.24,
+ "learning_rate": 0.0003213141025641026,
+ "loss": 0.618,
+ "step": 446
+ },
+ {
+ "epoch": 11.26,
+ "learning_rate": 0.00032091346153846156,
+ "loss": 0.6654,
+ "step": 447
+ },
+ {
+ "epoch": 11.29,
+ "learning_rate": 0.00032051282051282057,
+ "loss": 0.6632,
+ "step": 448
+ },
+ {
+ "epoch": 11.31,
+ "learning_rate": 0.00032011217948717946,
+ "loss": 0.6599,
+ "step": 449
+ },
+ {
+ "epoch": 11.34,
+ "learning_rate": 0.00031971153846153846,
+ "loss": 0.6537,
+ "step": 450
+ },
+ {
+ "epoch": 11.36,
+ "learning_rate": 0.0003193108974358974,
+ "loss": 0.678,
+ "step": 451
+ },
+ {
+ "epoch": 11.39,
+ "learning_rate": 0.0003189102564102564,
+ "loss": 0.684,
+ "step": 452
+ },
+ {
+ "epoch": 11.41,
+ "learning_rate": 0.00031850961538461536,
+ "loss": 0.6419,
+ "step": 453
+ },
+ {
+ "epoch": 11.44,
+ "learning_rate": 0.00031810897435897436,
+ "loss": 0.6614,
+ "step": 454
+ },
+ {
+ "epoch": 11.46,
+ "learning_rate": 0.0003177083333333333,
+ "loss": 0.7037,
+ "step": 455
+ },
+ {
+ "epoch": 11.49,
+ "learning_rate": 0.0003173076923076923,
+ "loss": 0.6944,
+ "step": 456
+ },
+ {
+ "epoch": 11.51,
+ "learning_rate": 0.00031690705128205126,
+ "loss": 0.6818,
+ "step": 457
+ },
+ {
+ "epoch": 11.54,
+ "learning_rate": 0.00031650641025641026,
+ "loss": 0.692,
+ "step": 458
+ },
+ {
+ "epoch": 11.57,
+ "learning_rate": 0.0003161057692307692,
+ "loss": 0.7028,
+ "step": 459
+ },
+ {
+ "epoch": 11.59,
+ "learning_rate": 0.0003157051282051282,
+ "loss": 0.7009,
+ "step": 460
+ },
+ {
+ "epoch": 11.62,
+ "learning_rate": 0.00031530448717948716,
+ "loss": 0.7325,
+ "step": 461
+ },
+ {
+ "epoch": 11.64,
+ "learning_rate": 0.00031490384615384616,
+ "loss": 0.6738,
+ "step": 462
+ },
+ {
+ "epoch": 11.67,
+ "learning_rate": 0.0003145032051282051,
+ "loss": 0.7008,
+ "step": 463
+ },
+ {
+ "epoch": 11.69,
+ "learning_rate": 0.0003141025641025641,
+ "loss": 0.6838,
+ "step": 464
+ },
+ {
+ "epoch": 11.72,
+ "learning_rate": 0.00031370192307692306,
+ "loss": 0.726,
+ "step": 465
+ },
+ {
+ "epoch": 11.74,
+ "learning_rate": 0.00031330128205128206,
+ "loss": 0.7167,
+ "step": 466
+ },
+ {
+ "epoch": 11.77,
+ "learning_rate": 0.000312900641025641,
+ "loss": 0.7236,
+ "step": 467
+ },
+ {
+ "epoch": 11.79,
+ "learning_rate": 0.0003125,
+ "loss": 0.7076,
+ "step": 468
+ },
+ {
+ "epoch": 11.82,
+ "learning_rate": 0.000312099358974359,
+ "loss": 0.7415,
+ "step": 469
+ },
+ {
+ "epoch": 11.84,
+ "learning_rate": 0.00031169871794871796,
+ "loss": 0.7313,
+ "step": 470
+ },
+ {
+ "epoch": 11.87,
+ "learning_rate": 0.00031129807692307696,
+ "loss": 0.7403,
+ "step": 471
+ },
+ {
+ "epoch": 11.89,
+ "learning_rate": 0.0003108974358974359,
+ "loss": 0.7145,
+ "step": 472
+ },
+ {
+ "epoch": 11.92,
+ "learning_rate": 0.0003104967948717949,
+ "loss": 0.7532,
+ "step": 473
+ },
+ {
+ "epoch": 11.94,
+ "learning_rate": 0.00031009615384615385,
+ "loss": 0.7326,
+ "step": 474
+ },
+ {
+ "epoch": 11.97,
+ "learning_rate": 0.00030969551282051286,
+ "loss": 0.7699,
+ "step": 475
+ },
+ {
+ "epoch": 11.99,
+ "learning_rate": 0.0003092948717948718,
+ "loss": 0.7251,
+ "step": 476
+ },
+ {
+ "epoch": 12.02,
+ "learning_rate": 0.0003088942307692308,
+ "loss": 0.5773,
+ "step": 477
+ },
+ {
+ "epoch": 12.04,
+ "learning_rate": 0.00030849358974358975,
+ "loss": 0.5288,
+ "step": 478
+ },
+ {
+ "epoch": 12.07,
+ "learning_rate": 0.00030809294871794876,
+ "loss": 0.5312,
+ "step": 479
+ },
+ {
+ "epoch": 12.09,
+ "learning_rate": 0.0003076923076923077,
+ "loss": 0.5445,
+ "step": 480
+ },
+ {
+ "epoch": 12.12,
+ "learning_rate": 0.0003072916666666667,
+ "loss": 0.5218,
+ "step": 481
+ },
+ {
+ "epoch": 12.14,
+ "learning_rate": 0.00030689102564102565,
+ "loss": 0.5311,
+ "step": 482
+ },
+ {
+ "epoch": 12.17,
+ "learning_rate": 0.00030649038461538465,
+ "loss": 0.5541,
+ "step": 483
+ },
+ {
+ "epoch": 12.2,
+ "learning_rate": 0.0003060897435897436,
+ "loss": 0.5242,
+ "step": 484
+ },
+ {
+ "epoch": 12.22,
+ "learning_rate": 0.0003056891025641026,
+ "loss": 0.5302,
+ "step": 485
+ },
+ {
+ "epoch": 12.25,
+ "learning_rate": 0.00030528846153846155,
+ "loss": 0.536,
+ "step": 486
+ },
+ {
+ "epoch": 12.27,
+ "learning_rate": 0.00030488782051282055,
+ "loss": 0.5545,
+ "step": 487
+ },
+ {
+ "epoch": 12.3,
+ "learning_rate": 0.00030448717948717945,
+ "loss": 0.5476,
+ "step": 488
+ },
+ {
+ "epoch": 12.32,
+ "learning_rate": 0.00030408653846153845,
+ "loss": 0.5528,
+ "step": 489
+ },
+ {
+ "epoch": 12.35,
+ "learning_rate": 0.0003036858974358974,
+ "loss": 0.5633,
+ "step": 490
+ },
+ {
+ "epoch": 12.37,
+ "learning_rate": 0.0003032852564102564,
+ "loss": 0.571,
+ "step": 491
+ },
+ {
+ "epoch": 12.4,
+ "learning_rate": 0.00030288461538461535,
+ "loss": 0.5637,
+ "step": 492
+ },
+ {
+ "epoch": 12.42,
+ "learning_rate": 0.00030248397435897435,
+ "loss": 0.5613,
+ "step": 493
+ },
+ {
+ "epoch": 12.45,
+ "learning_rate": 0.0003020833333333333,
+ "loss": 0.5999,
+ "step": 494
+ },
+ {
+ "epoch": 12.47,
+ "learning_rate": 0.0003016826923076923,
+ "loss": 0.5633,
+ "step": 495
+ },
+ {
+ "epoch": 12.5,
+ "learning_rate": 0.0003012820512820513,
+ "loss": 0.5811,
+ "step": 496
+ },
+ {
+ "epoch": 12.52,
+ "learning_rate": 0.00030088141025641025,
+ "loss": 0.5711,
+ "step": 497
+ },
+ {
+ "epoch": 12.55,
+ "learning_rate": 0.00030048076923076925,
+ "loss": 0.5669,
+ "step": 498
+ },
+ {
+ "epoch": 12.57,
+ "learning_rate": 0.0003000801282051282,
+ "loss": 0.5424,
+ "step": 499
+ },
+ {
+ "epoch": 12.6,
+ "learning_rate": 0.0002996794871794872,
+ "loss": 0.5592,
+ "step": 500
+ },
+ {
+ "epoch": 12.62,
+ "learning_rate": 0.00029927884615384614,
+ "loss": 0.5645,
+ "step": 501
+ },
+ {
+ "epoch": 12.65,
+ "learning_rate": 0.00029887820512820515,
+ "loss": 0.6089,
+ "step": 502
+ },
+ {
+ "epoch": 12.67,
+ "learning_rate": 0.0002984775641025641,
+ "loss": 0.57,
+ "step": 503
+ },
+ {
+ "epoch": 12.7,
+ "learning_rate": 0.0002980769230769231,
+ "loss": 0.57,
+ "step": 504
+ },
+ {
+ "epoch": 12.72,
+ "learning_rate": 0.00029767628205128204,
+ "loss": 0.5511,
+ "step": 505
+ },
+ {
+ "epoch": 12.75,
+ "learning_rate": 0.00029727564102564105,
+ "loss": 0.5765,
+ "step": 506
+ },
+ {
+ "epoch": 12.77,
+ "learning_rate": 0.000296875,
+ "loss": 0.5896,
+ "step": 507
+ },
+ {
+ "epoch": 12.8,
+ "learning_rate": 0.000296474358974359,
+ "loss": 0.6329,
+ "step": 508
+ },
+ {
+ "epoch": 12.83,
+ "learning_rate": 0.00029607371794871794,
+ "loss": 0.5682,
+ "step": 509
+ },
+ {
+ "epoch": 12.85,
+ "learning_rate": 0.00029567307692307694,
+ "loss": 0.58,
+ "step": 510
+ },
+ {
+ "epoch": 12.88,
+ "learning_rate": 0.0002952724358974359,
+ "loss": 0.6014,
+ "step": 511
+ },
+ {
+ "epoch": 12.9,
+ "learning_rate": 0.0002948717948717949,
+ "loss": 0.6082,
+ "step": 512
+ },
+ {
+ "epoch": 12.93,
+ "learning_rate": 0.00029447115384615384,
+ "loss": 0.6392,
+ "step": 513
+ },
+ {
+ "epoch": 12.95,
+ "learning_rate": 0.00029407051282051284,
+ "loss": 0.609,
+ "step": 514
+ },
+ {
+ "epoch": 12.98,
+ "learning_rate": 0.0002936698717948718,
+ "loss": 0.6247,
+ "step": 515
+ },
+ {
+ "epoch": 13.0,
+ "learning_rate": 0.0002932692307692308,
+ "loss": 0.5943,
+ "step": 516
+ },
+ {
+ "epoch": 13.03,
+ "learning_rate": 0.00029286858974358974,
+ "loss": 0.4499,
+ "step": 517
+ },
+ {
+ "epoch": 13.05,
+ "learning_rate": 0.00029246794871794874,
+ "loss": 0.4412,
+ "step": 518
+ },
+ {
+ "epoch": 13.08,
+ "learning_rate": 0.0002920673076923077,
+ "loss": 0.4409,
+ "step": 519
+ },
+ {
+ "epoch": 13.1,
+ "learning_rate": 0.0002916666666666667,
+ "loss": 0.4505,
+ "step": 520
+ },
+ {
+ "epoch": 13.13,
+ "learning_rate": 0.0002912660256410257,
+ "loss": 0.4489,
+ "step": 521
+ },
+ {
+ "epoch": 13.15,
+ "learning_rate": 0.00029086538461538464,
+ "loss": 0.4235,
+ "step": 522
+ },
+ {
+ "epoch": 13.18,
+ "learning_rate": 0.00029046474358974364,
+ "loss": 0.4276,
+ "step": 523
+ },
+ {
+ "epoch": 13.2,
+ "learning_rate": 0.0002900641025641026,
+ "loss": 0.4544,
+ "step": 524
+ },
+ {
+ "epoch": 13.23,
+ "learning_rate": 0.0002896634615384616,
+ "loss": 0.4625,
+ "step": 525
+ },
+ {
+ "epoch": 13.25,
+ "learning_rate": 0.00028926282051282054,
+ "loss": 0.4266,
+ "step": 526
+ },
+ {
+ "epoch": 13.28,
+ "learning_rate": 0.0002888621794871795,
+ "loss": 0.4459,
+ "step": 527
+ },
+ {
+ "epoch": 13.3,
+ "learning_rate": 0.00028846153846153843,
+ "loss": 0.4261,
+ "step": 528
+ },
+ {
+ "epoch": 13.33,
+ "learning_rate": 0.00028806089743589744,
+ "loss": 0.4376,
+ "step": 529
+ },
+ {
+ "epoch": 13.35,
+ "learning_rate": 0.0002876602564102564,
+ "loss": 0.4418,
+ "step": 530
+ },
+ {
+ "epoch": 13.38,
+ "learning_rate": 0.0002872596153846154,
+ "loss": 0.4667,
+ "step": 531
+ },
+ {
+ "epoch": 13.4,
+ "learning_rate": 0.00028685897435897433,
+ "loss": 0.4504,
+ "step": 532
+ },
+ {
+ "epoch": 13.43,
+ "learning_rate": 0.00028645833333333333,
+ "loss": 0.4691,
+ "step": 533
+ },
+ {
+ "epoch": 13.46,
+ "learning_rate": 0.0002860576923076923,
+ "loss": 0.4684,
+ "step": 534
+ },
+ {
+ "epoch": 13.48,
+ "learning_rate": 0.0002856570512820513,
+ "loss": 0.467,
+ "step": 535
+ },
+ {
+ "epoch": 13.51,
+ "learning_rate": 0.00028525641025641023,
+ "loss": 0.4716,
+ "step": 536
+ },
+ {
+ "epoch": 13.53,
+ "learning_rate": 0.00028485576923076923,
+ "loss": 0.4619,
+ "step": 537
+ },
+ {
+ "epoch": 13.56,
+ "learning_rate": 0.0002844551282051282,
+ "loss": 0.4775,
+ "step": 538
+ },
+ {
+ "epoch": 13.58,
+ "learning_rate": 0.0002840544871794872,
+ "loss": 0.4568,
+ "step": 539
+ },
+ {
+ "epoch": 13.61,
+ "learning_rate": 0.00028365384615384613,
+ "loss": 0.4763,
+ "step": 540
+ },
+ {
+ "epoch": 13.63,
+ "learning_rate": 0.00028325320512820513,
+ "loss": 0.466,
+ "step": 541
+ },
+ {
+ "epoch": 13.66,
+ "learning_rate": 0.0002828525641025641,
+ "loss": 0.4543,
+ "step": 542
+ },
+ {
+ "epoch": 13.68,
+ "learning_rate": 0.0002824519230769231,
+ "loss": 0.4666,
+ "step": 543
+ },
+ {
+ "epoch": 13.71,
+ "learning_rate": 0.00028205128205128203,
+ "loss": 0.489,
+ "step": 544
+ },
+ {
+ "epoch": 13.73,
+ "learning_rate": 0.00028165064102564103,
+ "loss": 0.491,
+ "step": 545
+ },
+ {
+ "epoch": 13.76,
+ "learning_rate": 0.00028125000000000003,
+ "loss": 0.4847,
+ "step": 546
+ },
+ {
+ "epoch": 13.78,
+ "learning_rate": 0.000280849358974359,
+ "loss": 0.4899,
+ "step": 547
+ },
+ {
+ "epoch": 13.81,
+ "learning_rate": 0.000280448717948718,
+ "loss": 0.5169,
+ "step": 548
+ },
+ {
+ "epoch": 13.83,
+ "learning_rate": 0.00028004807692307693,
+ "loss": 0.4968,
+ "step": 549
+ },
+ {
+ "epoch": 13.86,
+ "learning_rate": 0.00027964743589743593,
+ "loss": 0.4774,
+ "step": 550
+ },
+ {
+ "epoch": 13.88,
+ "learning_rate": 0.0002792467948717949,
+ "loss": 0.5105,
+ "step": 551
+ },
+ {
+ "epoch": 13.91,
+ "learning_rate": 0.0002788461538461539,
+ "loss": 0.4997,
+ "step": 552
+ },
+ {
+ "epoch": 13.93,
+ "learning_rate": 0.00027844551282051283,
+ "loss": 0.5088,
+ "step": 553
+ },
+ {
+ "epoch": 13.96,
+ "learning_rate": 0.00027804487179487183,
+ "loss": 0.5091,
+ "step": 554
+ },
+ {
+ "epoch": 13.98,
+ "learning_rate": 0.0002776442307692308,
+ "loss": 0.532,
+ "step": 555
+ },
+ {
+ "epoch": 14.01,
+ "learning_rate": 0.0002772435897435898,
+ "loss": 0.4487,
+ "step": 556
+ },
+ {
+ "epoch": 14.03,
+ "learning_rate": 0.00027684294871794873,
+ "loss": 0.3525,
+ "step": 557
+ },
+ {
+ "epoch": 14.06,
+ "learning_rate": 0.00027644230769230773,
+ "loss": 0.3715,
+ "step": 558
+ },
+ {
+ "epoch": 14.09,
+ "learning_rate": 0.0002760416666666667,
+ "loss": 0.3547,
+ "step": 559
+ },
+ {
+ "epoch": 14.11,
+ "learning_rate": 0.0002756410256410257,
+ "loss": 0.3862,
+ "step": 560
+ },
+ {
+ "epoch": 14.14,
+ "learning_rate": 0.0002752403846153846,
+ "loss": 0.3514,
+ "step": 561
+ },
+ {
+ "epoch": 14.16,
+ "learning_rate": 0.00027483974358974363,
+ "loss": 0.362,
+ "step": 562
+ },
+ {
+ "epoch": 14.19,
+ "learning_rate": 0.0002744391025641026,
+ "loss": 0.3457,
+ "step": 563
+ },
+ {
+ "epoch": 14.21,
+ "learning_rate": 0.0002740384615384616,
+ "loss": 0.3662,
+ "step": 564
+ },
+ {
+ "epoch": 14.24,
+ "learning_rate": 0.0002736378205128205,
+ "loss": 0.3746,
+ "step": 565
+ },
+ {
+ "epoch": 14.26,
+ "learning_rate": 0.00027323717948717947,
+ "loss": 0.3388,
+ "step": 566
+ },
+ {
+ "epoch": 14.29,
+ "learning_rate": 0.0002728365384615384,
+ "loss": 0.3824,
+ "step": 567
+ },
+ {
+ "epoch": 14.31,
+ "learning_rate": 0.0002724358974358974,
+ "loss": 0.3584,
+ "step": 568
+ },
+ {
+ "epoch": 14.34,
+ "learning_rate": 0.00027203525641025637,
+ "loss": 0.3544,
+ "step": 569
+ },
+ {
+ "epoch": 14.36,
+ "learning_rate": 0.00027163461538461537,
+ "loss": 0.3721,
+ "step": 570
+ },
+ {
+ "epoch": 14.39,
+ "learning_rate": 0.0002712339743589743,
+ "loss": 0.3769,
+ "step": 571
+ },
+ {
+ "epoch": 14.41,
+ "learning_rate": 0.0002708333333333333,
+ "loss": 0.3872,
+ "step": 572
+ },
+ {
+ "epoch": 14.44,
+ "learning_rate": 0.0002704326923076923,
+ "loss": 0.363,
+ "step": 573
+ },
+ {
+ "epoch": 14.46,
+ "learning_rate": 0.00027003205128205127,
+ "loss": 0.3842,
+ "step": 574
+ },
+ {
+ "epoch": 14.49,
+ "learning_rate": 0.00026963141025641027,
+ "loss": 0.3737,
+ "step": 575
+ },
+ {
+ "epoch": 14.51,
+ "learning_rate": 0.0002692307692307692,
+ "loss": 0.3981,
+ "step": 576
+ },
+ {
+ "epoch": 14.54,
+ "learning_rate": 0.0002688301282051282,
+ "loss": 0.3656,
+ "step": 577
+ },
+ {
+ "epoch": 14.56,
+ "learning_rate": 0.00026842948717948717,
+ "loss": 0.3852,
+ "step": 578
+ },
+ {
+ "epoch": 14.59,
+ "learning_rate": 0.00026802884615384617,
+ "loss": 0.3865,
+ "step": 579
+ },
+ {
+ "epoch": 14.61,
+ "learning_rate": 0.0002676282051282051,
+ "loss": 0.3747,
+ "step": 580
+ },
+ {
+ "epoch": 14.64,
+ "learning_rate": 0.0002672275641025641,
+ "loss": 0.3729,
+ "step": 581
+ },
+ {
+ "epoch": 14.66,
+ "learning_rate": 0.00026682692307692307,
+ "loss": 0.3785,
+ "step": 582
+ },
+ {
+ "epoch": 14.69,
+ "learning_rate": 0.00026642628205128207,
+ "loss": 0.4018,
+ "step": 583
+ },
+ {
+ "epoch": 14.71,
+ "learning_rate": 0.000266025641025641,
+ "loss": 0.3854,
+ "step": 584
+ },
+ {
+ "epoch": 14.74,
+ "learning_rate": 0.000265625,
+ "loss": 0.4161,
+ "step": 585
+ },
+ {
+ "epoch": 14.77,
+ "learning_rate": 0.00026522435897435897,
+ "loss": 0.4093,
+ "step": 586
+ },
+ {
+ "epoch": 14.79,
+ "learning_rate": 0.00026482371794871797,
+ "loss": 0.4126,
+ "step": 587
+ },
+ {
+ "epoch": 14.82,
+ "learning_rate": 0.0002644230769230769,
+ "loss": 0.3977,
+ "step": 588
+ },
+ {
+ "epoch": 14.84,
+ "learning_rate": 0.0002640224358974359,
+ "loss": 0.4286,
+ "step": 589
+ },
+ {
+ "epoch": 14.87,
+ "learning_rate": 0.00026362179487179487,
+ "loss": 0.4371,
+ "step": 590
+ },
+ {
+ "epoch": 14.89,
+ "learning_rate": 0.00026322115384615387,
+ "loss": 0.4066,
+ "step": 591
+ },
+ {
+ "epoch": 14.92,
+ "learning_rate": 0.0002628205128205128,
+ "loss": 0.433,
+ "step": 592
+ },
+ {
+ "epoch": 14.94,
+ "learning_rate": 0.0002624198717948718,
+ "loss": 0.4361,
+ "step": 593
+ },
+ {
+ "epoch": 14.97,
+ "learning_rate": 0.00026201923076923076,
+ "loss": 0.4276,
+ "step": 594
+ },
+ {
+ "epoch": 14.99,
+ "learning_rate": 0.00026161858974358977,
+ "loss": 0.4401,
+ "step": 595
+ },
+ {
+ "epoch": 15.02,
+ "learning_rate": 0.0002612179487179487,
+ "loss": 0.3514,
+ "step": 596
+ },
+ {
+ "epoch": 15.04,
+ "learning_rate": 0.0002608173076923077,
+ "loss": 0.306,
+ "step": 597
+ },
+ {
+ "epoch": 15.07,
+ "learning_rate": 0.0002604166666666667,
+ "loss": 0.2862,
+ "step": 598
+ },
+ {
+ "epoch": 15.09,
+ "learning_rate": 0.00026001602564102566,
+ "loss": 0.3042,
+ "step": 599
+ },
+ {
+ "epoch": 15.12,
+ "learning_rate": 0.00025961538461538467,
+ "loss": 0.2978,
+ "step": 600
+ },
+ {
+ "epoch": 15.14,
+ "learning_rate": 0.0002592147435897436,
+ "loss": 0.2975,
+ "step": 601
+ },
+ {
+ "epoch": 15.17,
+ "learning_rate": 0.0002588141025641026,
+ "loss": 0.3172,
+ "step": 602
+ },
+ {
+ "epoch": 15.19,
+ "learning_rate": 0.00025841346153846156,
+ "loss": 0.3046,
+ "step": 603
+ },
+ {
+ "epoch": 15.22,
+ "learning_rate": 0.00025801282051282057,
+ "loss": 0.3008,
+ "step": 604
+ },
+ {
+ "epoch": 15.24,
+ "learning_rate": 0.00025761217948717946,
+ "loss": 0.2956,
+ "step": 605
+ },
+ {
+ "epoch": 15.27,
+ "learning_rate": 0.00025721153846153846,
+ "loss": 0.3137,
+ "step": 606
+ },
+ {
+ "epoch": 15.29,
+ "learning_rate": 0.0002568108974358974,
+ "loss": 0.2961,
+ "step": 607
+ },
+ {
+ "epoch": 15.32,
+ "learning_rate": 0.0002564102564102564,
+ "loss": 0.3059,
+ "step": 608
+ },
+ {
+ "epoch": 15.34,
+ "learning_rate": 0.00025600961538461536,
+ "loss": 0.3077,
+ "step": 609
+ },
+ {
+ "epoch": 15.37,
+ "learning_rate": 0.00025560897435897436,
+ "loss": 0.3189,
+ "step": 610
+ },
+ {
+ "epoch": 15.4,
+ "learning_rate": 0.0002552083333333333,
+ "loss": 0.309,
+ "step": 611
+ },
+ {
+ "epoch": 15.42,
+ "learning_rate": 0.0002548076923076923,
+ "loss": 0.3324,
+ "step": 612
+ },
+ {
+ "epoch": 15.45,
+ "learning_rate": 0.00025440705128205126,
+ "loss": 0.3159,
+ "step": 613
+ },
+ {
+ "epoch": 15.47,
+ "learning_rate": 0.00025400641025641026,
+ "loss": 0.3188,
+ "step": 614
+ },
+ {
+ "epoch": 15.5,
+ "learning_rate": 0.0002536057692307692,
+ "loss": 0.3174,
+ "step": 615
+ },
+ {
+ "epoch": 15.52,
+ "learning_rate": 0.0002532051282051282,
+ "loss": 0.3074,
+ "step": 616
+ },
+ {
+ "epoch": 15.55,
+ "learning_rate": 0.00025280448717948716,
+ "loss": 0.3145,
+ "step": 617
+ },
+ {
+ "epoch": 15.57,
+ "learning_rate": 0.00025240384615384616,
+ "loss": 0.317,
+ "step": 618
+ },
+ {
+ "epoch": 15.6,
+ "learning_rate": 0.0002520032051282051,
+ "loss": 0.318,
+ "step": 619
+ },
+ {
+ "epoch": 15.62,
+ "learning_rate": 0.0002516025641025641,
+ "loss": 0.3236,
+ "step": 620
+ },
+ {
+ "epoch": 15.65,
+ "learning_rate": 0.00025120192307692305,
+ "loss": 0.3282,
+ "step": 621
+ },
+ {
+ "epoch": 15.67,
+ "learning_rate": 0.00025080128205128206,
+ "loss": 0.3394,
+ "step": 622
+ },
+ {
+ "epoch": 15.7,
+ "learning_rate": 0.000250400641025641,
+ "loss": 0.3278,
+ "step": 623
+ },
+ {
+ "epoch": 15.72,
+ "learning_rate": 0.00025,
+ "loss": 0.3255,
+ "step": 624
+ },
+ {
+ "epoch": 15.75,
+ "learning_rate": 0.000249599358974359,
+ "loss": 0.3256,
+ "step": 625
+ },
+ {
+ "epoch": 15.77,
+ "learning_rate": 0.00024919871794871795,
+ "loss": 0.3379,
+ "step": 626
+ },
+ {
+ "epoch": 15.8,
+ "learning_rate": 0.00024879807692307696,
+ "loss": 0.3384,
+ "step": 627
+ },
+ {
+ "epoch": 15.82,
+ "learning_rate": 0.0002483974358974359,
+ "loss": 0.3483,
+ "step": 628
+ },
+ {
+ "epoch": 15.85,
+ "learning_rate": 0.0002479967948717949,
+ "loss": 0.3478,
+ "step": 629
+ },
+ {
+ "epoch": 15.87,
+ "learning_rate": 0.00024759615384615385,
+ "loss": 0.337,
+ "step": 630
+ },
+ {
+ "epoch": 15.9,
+ "learning_rate": 0.00024719551282051286,
+ "loss": 0.3287,
+ "step": 631
+ },
+ {
+ "epoch": 15.92,
+ "learning_rate": 0.0002467948717948718,
+ "loss": 0.3478,
+ "step": 632
+ },
+ {
+ "epoch": 15.95,
+ "learning_rate": 0.0002463942307692308,
+ "loss": 0.3567,
+ "step": 633
+ },
+ {
+ "epoch": 15.97,
+ "learning_rate": 0.00024599358974358975,
+ "loss": 0.3359,
+ "step": 634
+ },
+ {
+ "epoch": 16.0,
+ "learning_rate": 0.0002455929487179487,
+ "loss": 0.3492,
+ "step": 635
+ },
+ {
+ "epoch": 16.03,
+ "learning_rate": 0.0002451923076923077,
+ "loss": 0.2497,
+ "step": 636
+ },
+ {
+ "epoch": 16.05,
+ "learning_rate": 0.00024479166666666665,
+ "loss": 0.2416,
+ "step": 637
+ },
+ {
+ "epoch": 16.08,
+ "learning_rate": 0.00024439102564102565,
+ "loss": 0.2576,
+ "step": 638
+ },
+ {
+ "epoch": 16.1,
+ "learning_rate": 0.00024399038461538463,
+ "loss": 0.243,
+ "step": 639
+ },
+ {
+ "epoch": 16.13,
+ "learning_rate": 0.0002435897435897436,
+ "loss": 0.2527,
+ "step": 640
+ },
+ {
+ "epoch": 16.15,
+ "learning_rate": 0.00024318910256410258,
+ "loss": 0.2606,
+ "step": 641
+ },
+ {
+ "epoch": 16.18,
+ "learning_rate": 0.00024278846153846155,
+ "loss": 0.2555,
+ "step": 642
+ },
+ {
+ "epoch": 16.2,
+ "learning_rate": 0.00024238782051282052,
+ "loss": 0.2581,
+ "step": 643
+ },
+ {
+ "epoch": 16.23,
+ "learning_rate": 0.0002419871794871795,
+ "loss": 0.2508,
+ "step": 644
+ },
+ {
+ "epoch": 16.25,
+ "learning_rate": 0.00024158653846153847,
+ "loss": 0.281,
+ "step": 645
+ },
+ {
+ "epoch": 16.28,
+ "learning_rate": 0.00024118589743589745,
+ "loss": 0.2626,
+ "step": 646
+ },
+ {
+ "epoch": 16.3,
+ "learning_rate": 0.00024078525641025642,
+ "loss": 0.2668,
+ "step": 647
+ },
+ {
+ "epoch": 16.33,
+ "learning_rate": 0.0002403846153846154,
+ "loss": 0.2675,
+ "step": 648
+ },
+ {
+ "epoch": 16.35,
+ "learning_rate": 0.00023998397435897437,
+ "loss": 0.2491,
+ "step": 649
+ },
+ {
+ "epoch": 16.38,
+ "learning_rate": 0.00023958333333333335,
+ "loss": 0.2459,
+ "step": 650
+ },
+ {
+ "epoch": 16.4,
+ "learning_rate": 0.00023918269230769232,
+ "loss": 0.2515,
+ "step": 651
+ },
+ {
+ "epoch": 16.43,
+ "learning_rate": 0.0002387820512820513,
+ "loss": 0.2612,
+ "step": 652
+ },
+ {
+ "epoch": 16.45,
+ "learning_rate": 0.00023838141025641027,
+ "loss": 0.2612,
+ "step": 653
+ },
+ {
+ "epoch": 16.48,
+ "learning_rate": 0.00023798076923076922,
+ "loss": 0.265,
+ "step": 654
+ },
+ {
+ "epoch": 16.5,
+ "learning_rate": 0.0002375801282051282,
+ "loss": 0.2634,
+ "step": 655
+ },
+ {
+ "epoch": 16.53,
+ "learning_rate": 0.00023717948717948717,
+ "loss": 0.2839,
+ "step": 656
+ },
+ {
+ "epoch": 16.55,
+ "learning_rate": 0.00023677884615384614,
+ "loss": 0.2595,
+ "step": 657
+ },
+ {
+ "epoch": 16.58,
+ "learning_rate": 0.00023637820512820512,
+ "loss": 0.2764,
+ "step": 658
+ },
+ {
+ "epoch": 16.6,
+ "learning_rate": 0.0002359775641025641,
+ "loss": 0.2553,
+ "step": 659
+ },
+ {
+ "epoch": 16.63,
+ "learning_rate": 0.00023557692307692307,
+ "loss": 0.2716,
+ "step": 660
+ },
+ {
+ "epoch": 16.66,
+ "learning_rate": 0.00023517628205128204,
+ "loss": 0.289,
+ "step": 661
+ },
+ {
+ "epoch": 16.68,
+ "learning_rate": 0.00023477564102564102,
+ "loss": 0.2483,
+ "step": 662
+ },
+ {
+ "epoch": 16.71,
+ "learning_rate": 0.000234375,
+ "loss": 0.269,
+ "step": 663
+ },
+ {
+ "epoch": 16.73,
+ "learning_rate": 0.000233974358974359,
+ "loss": 0.2738,
+ "step": 664
+ },
+ {
+ "epoch": 16.76,
+ "learning_rate": 0.00023357371794871797,
+ "loss": 0.2646,
+ "step": 665
+ },
+ {
+ "epoch": 16.78,
+ "learning_rate": 0.00023317307692307694,
+ "loss": 0.2793,
+ "step": 666
+ },
+ {
+ "epoch": 16.81,
+ "learning_rate": 0.00023277243589743592,
+ "loss": 0.2666,
+ "step": 667
+ },
+ {
+ "epoch": 16.83,
+ "learning_rate": 0.0002323717948717949,
+ "loss": 0.2799,
+ "step": 668
+ },
+ {
+ "epoch": 16.86,
+ "learning_rate": 0.00023197115384615387,
+ "loss": 0.2677,
+ "step": 669
+ },
+ {
+ "epoch": 16.88,
+ "learning_rate": 0.00023157051282051284,
+ "loss": 0.2889,
+ "step": 670
+ },
+ {
+ "epoch": 16.91,
+ "learning_rate": 0.00023116987179487182,
+ "loss": 0.2869,
+ "step": 671
+ },
+ {
+ "epoch": 16.93,
+ "learning_rate": 0.0002307692307692308,
+ "loss": 0.2814,
+ "step": 672
+ },
+ {
+ "epoch": 16.96,
+ "learning_rate": 0.00023036858974358974,
+ "loss": 0.2881,
+ "step": 673
+ },
+ {
+ "epoch": 16.98,
+ "learning_rate": 0.0002299679487179487,
+ "loss": 0.2777,
+ "step": 674
+ },
+ {
+ "epoch": 17.01,
+ "learning_rate": 0.0002295673076923077,
+ "loss": 0.2713,
+ "step": 675
+ },
+ {
+ "epoch": 17.03,
+ "learning_rate": 0.00022916666666666666,
+ "loss": 0.2176,
+ "step": 676
+ },
+ {
+ "epoch": 17.06,
+ "learning_rate": 0.00022876602564102564,
+ "loss": 0.2032,
+ "step": 677
+ },
+ {
+ "epoch": 17.08,
+ "learning_rate": 0.0002283653846153846,
+ "loss": 0.2207,
+ "step": 678
+ },
+ {
+ "epoch": 17.11,
+ "learning_rate": 0.0002279647435897436,
+ "loss": 0.2183,
+ "step": 679
+ },
+ {
+ "epoch": 17.13,
+ "learning_rate": 0.00022756410256410256,
+ "loss": 0.2238,
+ "step": 680
+ },
+ {
+ "epoch": 17.16,
+ "learning_rate": 0.00022716346153846154,
+ "loss": 0.2126,
+ "step": 681
+ },
+ {
+ "epoch": 17.18,
+ "learning_rate": 0.0002267628205128205,
+ "loss": 0.2104,
+ "step": 682
+ },
+ {
+ "epoch": 17.21,
+ "learning_rate": 0.00022636217948717949,
+ "loss": 0.2151,
+ "step": 683
+ },
+ {
+ "epoch": 17.23,
+ "learning_rate": 0.00022596153846153846,
+ "loss": 0.2093,
+ "step": 684
+ },
+ {
+ "epoch": 17.26,
+ "learning_rate": 0.00022556089743589743,
+ "loss": 0.2308,
+ "step": 685
+ },
+ {
+ "epoch": 17.29,
+ "learning_rate": 0.0002251602564102564,
+ "loss": 0.2276,
+ "step": 686
+ },
+ {
+ "epoch": 17.31,
+ "learning_rate": 0.00022475961538461538,
+ "loss": 0.2328,
+ "step": 687
+ },
+ {
+ "epoch": 17.34,
+ "learning_rate": 0.00022435897435897436,
+ "loss": 0.2076,
+ "step": 688
+ },
+ {
+ "epoch": 17.36,
+ "learning_rate": 0.00022395833333333336,
+ "loss": 0.2125,
+ "step": 689
+ },
+ {
+ "epoch": 17.39,
+ "learning_rate": 0.00022355769230769234,
+ "loss": 0.2181,
+ "step": 690
+ },
+ {
+ "epoch": 17.41,
+ "learning_rate": 0.0002231570512820513,
+ "loss": 0.2153,
+ "step": 691
+ },
+ {
+ "epoch": 17.44,
+ "learning_rate": 0.00022275641025641028,
+ "loss": 0.2122,
+ "step": 692
+ },
+ {
+ "epoch": 17.46,
+ "learning_rate": 0.00022235576923076923,
+ "loss": 0.2262,
+ "step": 693
+ },
+ {
+ "epoch": 17.49,
+ "learning_rate": 0.0002219551282051282,
+ "loss": 0.2206,
+ "step": 694
+ },
+ {
+ "epoch": 17.51,
+ "learning_rate": 0.00022155448717948718,
+ "loss": 0.2223,
+ "step": 695
+ },
+ {
+ "epoch": 17.54,
+ "learning_rate": 0.00022115384615384616,
+ "loss": 0.2306,
+ "step": 696
+ },
+ {
+ "epoch": 17.56,
+ "learning_rate": 0.00022075320512820513,
+ "loss": 0.213,
+ "step": 697
+ },
+ {
+ "epoch": 17.59,
+ "learning_rate": 0.0002203525641025641,
+ "loss": 0.2198,
+ "step": 698
+ },
+ {
+ "epoch": 17.61,
+ "learning_rate": 0.00021995192307692308,
+ "loss": 0.2299,
+ "step": 699
+ },
+ {
+ "epoch": 17.64,
+ "learning_rate": 0.00021955128205128206,
+ "loss": 0.2213,
+ "step": 700
+ },
+ {
+ "epoch": 17.66,
+ "learning_rate": 0.00021915064102564103,
+ "loss": 0.2183,
+ "step": 701
+ },
+ {
+ "epoch": 17.69,
+ "learning_rate": 0.00021875,
+ "loss": 0.2301,
+ "step": 702
+ },
+ {
+ "epoch": 17.71,
+ "learning_rate": 0.00021834935897435898,
+ "loss": 0.2348,
+ "step": 703
+ },
+ {
+ "epoch": 17.74,
+ "learning_rate": 0.00021794871794871795,
+ "loss": 0.229,
+ "step": 704
+ },
+ {
+ "epoch": 17.76,
+ "learning_rate": 0.00021754807692307693,
+ "loss": 0.2397,
+ "step": 705
+ },
+ {
+ "epoch": 17.79,
+ "learning_rate": 0.0002171474358974359,
+ "loss": 0.2254,
+ "step": 706
+ },
+ {
+ "epoch": 17.81,
+ "learning_rate": 0.00021674679487179488,
+ "loss": 0.2306,
+ "step": 707
+ },
+ {
+ "epoch": 17.84,
+ "learning_rate": 0.00021634615384615385,
+ "loss": 0.2388,
+ "step": 708
+ },
+ {
+ "epoch": 17.86,
+ "learning_rate": 0.00021594551282051283,
+ "loss": 0.2374,
+ "step": 709
+ },
+ {
+ "epoch": 17.89,
+ "learning_rate": 0.0002155448717948718,
+ "loss": 0.2412,
+ "step": 710
+ },
+ {
+ "epoch": 17.91,
+ "learning_rate": 0.00021514423076923078,
+ "loss": 0.2203,
+ "step": 711
+ },
+ {
+ "epoch": 17.94,
+ "learning_rate": 0.00021474358974358972,
+ "loss": 0.2359,
+ "step": 712
+ },
+ {
+ "epoch": 17.97,
+ "learning_rate": 0.0002143429487179487,
+ "loss": 0.2343,
+ "step": 713
+ },
+ {
+ "epoch": 17.99,
+ "learning_rate": 0.00021394230769230767,
+ "loss": 0.2397,
+ "step": 714
+ },
+ {
+ "epoch": 18.02,
+ "learning_rate": 0.00021354166666666665,
+ "loss": 0.2055,
+ "step": 715
+ },
+ {
+ "epoch": 18.04,
+ "learning_rate": 0.00021314102564102565,
+ "loss": 0.1723,
+ "step": 716
+ },
+ {
+ "epoch": 18.07,
+ "learning_rate": 0.00021274038461538463,
+ "loss": 0.1782,
+ "step": 717
+ },
+ {
+ "epoch": 18.09,
+ "learning_rate": 0.0002123397435897436,
+ "loss": 0.1813,
+ "step": 718
+ },
+ {
+ "epoch": 18.12,
+ "learning_rate": 0.00021193910256410257,
+ "loss": 0.1834,
+ "step": 719
+ },
+ {
+ "epoch": 18.14,
+ "learning_rate": 0.00021153846153846155,
+ "loss": 0.1759,
+ "step": 720
+ },
+ {
+ "epoch": 18.17,
+ "learning_rate": 0.00021113782051282052,
+ "loss": 0.1855,
+ "step": 721
+ },
+ {
+ "epoch": 18.19,
+ "learning_rate": 0.0002107371794871795,
+ "loss": 0.1756,
+ "step": 722
+ },
+ {
+ "epoch": 18.22,
+ "learning_rate": 0.00021033653846153847,
+ "loss": 0.1783,
+ "step": 723
+ },
+ {
+ "epoch": 18.24,
+ "learning_rate": 0.00020993589743589745,
+ "loss": 0.1899,
+ "step": 724
+ },
+ {
+ "epoch": 18.27,
+ "learning_rate": 0.00020953525641025642,
+ "loss": 0.1817,
+ "step": 725
+ },
+ {
+ "epoch": 18.29,
+ "learning_rate": 0.0002091346153846154,
+ "loss": 0.1856,
+ "step": 726
+ },
+ {
+ "epoch": 18.32,
+ "learning_rate": 0.00020873397435897437,
+ "loss": 0.1849,
+ "step": 727
+ },
+ {
+ "epoch": 18.34,
+ "learning_rate": 0.00020833333333333335,
+ "loss": 0.1758,
+ "step": 728
+ },
+ {
+ "epoch": 18.37,
+ "learning_rate": 0.00020793269230769232,
+ "loss": 0.1907,
+ "step": 729
+ },
+ {
+ "epoch": 18.39,
+ "learning_rate": 0.0002075320512820513,
+ "loss": 0.1836,
+ "step": 730
+ },
+ {
+ "epoch": 18.42,
+ "learning_rate": 0.00020713141025641027,
+ "loss": 0.1828,
+ "step": 731
+ },
+ {
+ "epoch": 18.44,
+ "learning_rate": 0.00020673076923076922,
+ "loss": 0.1916,
+ "step": 732
+ },
+ {
+ "epoch": 18.47,
+ "learning_rate": 0.0002063301282051282,
+ "loss": 0.1849,
+ "step": 733
+ },
+ {
+ "epoch": 18.49,
+ "learning_rate": 0.00020592948717948717,
+ "loss": 0.1929,
+ "step": 734
+ },
+ {
+ "epoch": 18.52,
+ "learning_rate": 0.00020552884615384614,
+ "loss": 0.1779,
+ "step": 735
+ },
+ {
+ "epoch": 18.54,
+ "learning_rate": 0.00020512820512820512,
+ "loss": 0.1805,
+ "step": 736
+ },
+ {
+ "epoch": 18.57,
+ "learning_rate": 0.0002047275641025641,
+ "loss": 0.1891,
+ "step": 737
+ },
+ {
+ "epoch": 18.6,
+ "learning_rate": 0.00020432692307692307,
+ "loss": 0.1932,
+ "step": 738
+ },
+ {
+ "epoch": 18.62,
+ "learning_rate": 0.00020392628205128204,
+ "loss": 0.1923,
+ "step": 739
+ },
+ {
+ "epoch": 18.65,
+ "learning_rate": 0.00020352564102564102,
+ "loss": 0.1882,
+ "step": 740
+ },
+ {
+ "epoch": 18.67,
+ "learning_rate": 0.00020312500000000002,
+ "loss": 0.2012,
+ "step": 741
+ },
+ {
+ "epoch": 18.7,
+ "learning_rate": 0.000202724358974359,
+ "loss": 0.1963,
+ "step": 742
+ },
+ {
+ "epoch": 18.72,
+ "learning_rate": 0.00020232371794871797,
+ "loss": 0.202,
+ "step": 743
+ },
+ {
+ "epoch": 18.75,
+ "learning_rate": 0.00020192307692307694,
+ "loss": 0.1996,
+ "step": 744
+ },
+ {
+ "epoch": 18.77,
+ "learning_rate": 0.00020152243589743592,
+ "loss": 0.2056,
+ "step": 745
+ },
+ {
+ "epoch": 18.8,
+ "learning_rate": 0.0002011217948717949,
+ "loss": 0.1946,
+ "step": 746
+ },
+ {
+ "epoch": 18.82,
+ "learning_rate": 0.00020072115384615387,
+ "loss": 0.1954,
+ "step": 747
+ },
+ {
+ "epoch": 18.85,
+ "learning_rate": 0.00020032051282051284,
+ "loss": 0.1966,
+ "step": 748
+ },
+ {
+ "epoch": 18.87,
+ "learning_rate": 0.00019991987179487182,
+ "loss": 0.2077,
+ "step": 749
+ },
+ {
+ "epoch": 18.9,
+ "learning_rate": 0.0001995192307692308,
+ "loss": 0.2,
+ "step": 750
+ },
+ {
+ "epoch": 18.92,
+ "learning_rate": 0.00019911858974358974,
+ "loss": 0.2092,
+ "step": 751
+ },
+ {
+ "epoch": 18.95,
+ "learning_rate": 0.0001987179487179487,
+ "loss": 0.2023,
+ "step": 752
+ },
+ {
+ "epoch": 18.97,
+ "learning_rate": 0.0001983173076923077,
+ "loss": 0.2056,
+ "step": 753
+ },
+ {
+ "epoch": 19.0,
+ "learning_rate": 0.00019791666666666666,
+ "loss": 0.2059,
+ "step": 754
+ },
+ {
+ "epoch": 19.02,
+ "learning_rate": 0.00019751602564102564,
+ "loss": 0.1581,
+ "step": 755
+ },
+ {
+ "epoch": 19.05,
+ "learning_rate": 0.0001971153846153846,
+ "loss": 0.155,
+ "step": 756
+ },
+ {
+ "epoch": 19.07,
+ "learning_rate": 0.00019671474358974359,
+ "loss": 0.156,
+ "step": 757
+ },
+ {
+ "epoch": 19.1,
+ "learning_rate": 0.00019631410256410256,
+ "loss": 0.161,
+ "step": 758
+ },
+ {
+ "epoch": 19.12,
+ "learning_rate": 0.00019591346153846154,
+ "loss": 0.1608,
+ "step": 759
+ },
+ {
+ "epoch": 19.15,
+ "learning_rate": 0.0001955128205128205,
+ "loss": 0.1576,
+ "step": 760
+ },
+ {
+ "epoch": 19.17,
+ "learning_rate": 0.00019511217948717948,
+ "loss": 0.1539,
+ "step": 761
+ },
+ {
+ "epoch": 19.2,
+ "learning_rate": 0.00019471153846153846,
+ "loss": 0.1549,
+ "step": 762
+ },
+ {
+ "epoch": 19.23,
+ "learning_rate": 0.00019431089743589743,
+ "loss": 0.1593,
+ "step": 763
+ },
+ {
+ "epoch": 19.25,
+ "learning_rate": 0.0001939102564102564,
+ "loss": 0.162,
+ "step": 764
+ },
+ {
+ "epoch": 19.28,
+ "learning_rate": 0.00019350961538461538,
+ "loss": 0.1591,
+ "step": 765
+ },
+ {
+ "epoch": 19.3,
+ "learning_rate": 0.00019310897435897436,
+ "loss": 0.1569,
+ "step": 766
+ },
+ {
+ "epoch": 19.33,
+ "learning_rate": 0.00019270833333333336,
+ "loss": 0.1628,
+ "step": 767
+ },
+ {
+ "epoch": 19.35,
+ "learning_rate": 0.00019230769230769233,
+ "loss": 0.1529,
+ "step": 768
+ },
+ {
+ "epoch": 19.38,
+ "learning_rate": 0.0001919070512820513,
+ "loss": 0.1606,
+ "step": 769
+ },
+ {
+ "epoch": 19.4,
+ "learning_rate": 0.00019150641025641028,
+ "loss": 0.1564,
+ "step": 770
+ },
+ {
+ "epoch": 19.43,
+ "learning_rate": 0.00019110576923076923,
+ "loss": 0.1599,
+ "step": 771
+ },
+ {
+ "epoch": 19.45,
+ "learning_rate": 0.0001907051282051282,
+ "loss": 0.1653,
+ "step": 772
+ },
+ {
+ "epoch": 19.48,
+ "learning_rate": 0.00019030448717948718,
+ "loss": 0.1587,
+ "step": 773
+ },
+ {
+ "epoch": 19.5,
+ "learning_rate": 0.00018990384615384616,
+ "loss": 0.1683,
+ "step": 774
+ },
+ {
+ "epoch": 19.53,
+ "learning_rate": 0.00018950320512820513,
+ "loss": 0.157,
+ "step": 775
+ },
+ {
+ "epoch": 19.55,
+ "learning_rate": 0.0001891025641025641,
+ "loss": 0.1588,
+ "step": 776
+ },
+ {
+ "epoch": 19.58,
+ "learning_rate": 0.00018870192307692308,
+ "loss": 0.1709,
+ "step": 777
+ },
+ {
+ "epoch": 19.6,
+ "learning_rate": 0.00018830128205128205,
+ "loss": 0.1635,
+ "step": 778
+ },
+ {
+ "epoch": 19.63,
+ "learning_rate": 0.00018790064102564103,
+ "loss": 0.1581,
+ "step": 779
+ },
+ {
+ "epoch": 19.65,
+ "learning_rate": 0.0001875,
+ "loss": 0.1664,
+ "step": 780
+ },
+ {
+ "epoch": 19.68,
+ "learning_rate": 0.00018709935897435898,
+ "loss": 0.1712,
+ "step": 781
+ },
+ {
+ "epoch": 19.7,
+ "learning_rate": 0.00018669871794871795,
+ "loss": 0.165,
+ "step": 782
+ },
+ {
+ "epoch": 19.73,
+ "learning_rate": 0.00018629807692307693,
+ "loss": 0.1692,
+ "step": 783
+ },
+ {
+ "epoch": 19.75,
+ "learning_rate": 0.0001858974358974359,
+ "loss": 0.1658,
+ "step": 784
+ },
+ {
+ "epoch": 19.78,
+ "learning_rate": 0.00018549679487179488,
+ "loss": 0.1702,
+ "step": 785
+ },
+ {
+ "epoch": 19.8,
+ "learning_rate": 0.00018509615384615385,
+ "loss": 0.1672,
+ "step": 786
+ },
+ {
+ "epoch": 19.83,
+ "learning_rate": 0.00018469551282051283,
+ "loss": 0.1708,
+ "step": 787
+ },
+ {
+ "epoch": 19.86,
+ "learning_rate": 0.0001842948717948718,
+ "loss": 0.1705,
+ "step": 788
+ },
+ {
+ "epoch": 19.88,
+ "learning_rate": 0.00018389423076923078,
+ "loss": 0.1853,
+ "step": 789
+ },
+ {
+ "epoch": 19.91,
+ "learning_rate": 0.00018349358974358972,
+ "loss": 0.1701,
+ "step": 790
+ },
+ {
+ "epoch": 19.93,
+ "learning_rate": 0.0001830929487179487,
+ "loss": 0.1726,
+ "step": 791
+ },
+ {
+ "epoch": 19.96,
+ "learning_rate": 0.00018269230769230767,
+ "loss": 0.1829,
+ "step": 792
+ },
+ {
+ "epoch": 19.98,
+ "learning_rate": 0.00018229166666666665,
+ "loss": 0.1749,
+ "step": 793
+ },
+ {
+ "epoch": 20.01,
+ "learning_rate": 0.00018189102564102565,
+ "loss": 0.1559,
+ "step": 794
+ },
+ {
+ "epoch": 20.03,
+ "learning_rate": 0.00018149038461538462,
+ "loss": 0.1424,
+ "step": 795
+ },
+ {
+ "epoch": 20.06,
+ "learning_rate": 0.0001810897435897436,
+ "loss": 0.1386,
+ "step": 796
+ },
+ {
+ "epoch": 20.08,
+ "learning_rate": 0.00018068910256410257,
+ "loss": 0.1388,
+ "step": 797
+ },
+ {
+ "epoch": 20.11,
+ "learning_rate": 0.00018028846153846155,
+ "loss": 0.1385,
+ "step": 798
+ },
+ {
+ "epoch": 20.13,
+ "learning_rate": 0.00017988782051282052,
+ "loss": 0.1349,
+ "step": 799
+ },
+ {
+ "epoch": 20.16,
+ "learning_rate": 0.0001794871794871795,
+ "loss": 0.1357,
+ "step": 800
+ },
+ {
+ "epoch": 20.18,
+ "learning_rate": 0.00017908653846153847,
+ "loss": 0.1384,
+ "step": 801
+ },
+ {
+ "epoch": 20.21,
+ "learning_rate": 0.00017868589743589745,
+ "loss": 0.1425,
+ "step": 802
+ },
+ {
+ "epoch": 20.23,
+ "learning_rate": 0.00017828525641025642,
+ "loss": 0.1337,
+ "step": 803
+ },
+ {
+ "epoch": 20.26,
+ "learning_rate": 0.0001778846153846154,
+ "loss": 0.136,
+ "step": 804
+ },
+ {
+ "epoch": 20.28,
+ "learning_rate": 0.00017748397435897437,
+ "loss": 0.1382,
+ "step": 805
+ },
+ {
+ "epoch": 20.31,
+ "learning_rate": 0.00017708333333333335,
+ "loss": 0.1396,
+ "step": 806
+ },
+ {
+ "epoch": 20.33,
+ "learning_rate": 0.00017668269230769232,
+ "loss": 0.1411,
+ "step": 807
+ },
+ {
+ "epoch": 20.36,
+ "learning_rate": 0.0001762820512820513,
+ "loss": 0.1375,
+ "step": 808
+ },
+ {
+ "epoch": 20.38,
+ "learning_rate": 0.00017588141025641027,
+ "loss": 0.1457,
+ "step": 809
+ },
+ {
+ "epoch": 20.41,
+ "learning_rate": 0.00017548076923076922,
+ "loss": 0.1447,
+ "step": 810
+ },
+ {
+ "epoch": 20.43,
+ "learning_rate": 0.0001750801282051282,
+ "loss": 0.1381,
+ "step": 811
+ },
+ {
+ "epoch": 20.46,
+ "learning_rate": 0.00017467948717948717,
+ "loss": 0.1443,
+ "step": 812
+ },
+ {
+ "epoch": 20.49,
+ "learning_rate": 0.00017427884615384614,
+ "loss": 0.1427,
+ "step": 813
+ },
+ {
+ "epoch": 20.51,
+ "learning_rate": 0.00017387820512820512,
+ "loss": 0.1425,
+ "step": 814
+ },
+ {
+ "epoch": 20.54,
+ "learning_rate": 0.0001734775641025641,
+ "loss": 0.1428,
+ "step": 815
+ },
+ {
+ "epoch": 20.56,
+ "learning_rate": 0.00017307692307692307,
+ "loss": 0.1454,
+ "step": 816
+ },
+ {
+ "epoch": 20.59,
+ "learning_rate": 0.00017267628205128204,
+ "loss": 0.1422,
+ "step": 817
+ },
+ {
+ "epoch": 20.61,
+ "learning_rate": 0.00017227564102564102,
+ "loss": 0.1426,
+ "step": 818
+ },
+ {
+ "epoch": 20.64,
+ "learning_rate": 0.000171875,
+ "loss": 0.1453,
+ "step": 819
+ },
+ {
+ "epoch": 20.66,
+ "learning_rate": 0.000171474358974359,
+ "loss": 0.1472,
+ "step": 820
+ },
+ {
+ "epoch": 20.69,
+ "learning_rate": 0.00017107371794871797,
+ "loss": 0.1479,
+ "step": 821
+ },
+ {
+ "epoch": 20.71,
+ "learning_rate": 0.00017067307692307694,
+ "loss": 0.1387,
+ "step": 822
+ },
+ {
+ "epoch": 20.74,
+ "learning_rate": 0.00017027243589743592,
+ "loss": 0.15,
+ "step": 823
+ },
+ {
+ "epoch": 20.76,
+ "learning_rate": 0.0001698717948717949,
+ "loss": 0.1429,
+ "step": 824
+ },
+ {
+ "epoch": 20.79,
+ "learning_rate": 0.00016947115384615387,
+ "loss": 0.1532,
+ "step": 825
+ },
+ {
+ "epoch": 20.81,
+ "learning_rate": 0.00016907051282051284,
+ "loss": 0.1452,
+ "step": 826
+ },
+ {
+ "epoch": 20.84,
+ "learning_rate": 0.00016866987179487181,
+ "loss": 0.1445,
+ "step": 827
+ },
+ {
+ "epoch": 20.86,
+ "learning_rate": 0.0001682692307692308,
+ "loss": 0.1461,
+ "step": 828
+ },
+ {
+ "epoch": 20.89,
+ "learning_rate": 0.00016786858974358974,
+ "loss": 0.1544,
+ "step": 829
+ },
+ {
+ "epoch": 20.91,
+ "learning_rate": 0.0001674679487179487,
+ "loss": 0.1476,
+ "step": 830
+ },
+ {
+ "epoch": 20.94,
+ "learning_rate": 0.00016706730769230769,
+ "loss": 0.1535,
+ "step": 831
+ },
+ {
+ "epoch": 20.96,
+ "learning_rate": 0.00016666666666666666,
+ "loss": 0.1498,
+ "step": 832
+ },
+ {
+ "epoch": 20.99,
+ "learning_rate": 0.00016626602564102564,
+ "loss": 0.1514,
+ "step": 833
+ },
+ {
+ "epoch": 21.01,
+ "learning_rate": 0.0001658653846153846,
+ "loss": 0.1405,
+ "step": 834
+ },
+ {
+ "epoch": 21.04,
+ "learning_rate": 0.00016546474358974359,
+ "loss": 0.1218,
+ "step": 835
+ },
+ {
+ "epoch": 21.06,
+ "learning_rate": 0.00016506410256410256,
+ "loss": 0.1228,
+ "step": 836
+ },
+ {
+ "epoch": 21.09,
+ "learning_rate": 0.00016466346153846153,
+ "loss": 0.119,
+ "step": 837
+ },
+ {
+ "epoch": 21.11,
+ "learning_rate": 0.0001642628205128205,
+ "loss": 0.1278,
+ "step": 838
+ },
+ {
+ "epoch": 21.14,
+ "learning_rate": 0.00016386217948717948,
+ "loss": 0.1205,
+ "step": 839
+ },
+ {
+ "epoch": 21.17,
+ "learning_rate": 0.00016346153846153846,
+ "loss": 0.1234,
+ "step": 840
+ },
+ {
+ "epoch": 21.19,
+ "learning_rate": 0.00016306089743589743,
+ "loss": 0.1226,
+ "step": 841
+ },
+ {
+ "epoch": 21.22,
+ "learning_rate": 0.0001626602564102564,
+ "loss": 0.1216,
+ "step": 842
+ },
+ {
+ "epoch": 21.24,
+ "learning_rate": 0.00016225961538461538,
+ "loss": 0.1221,
+ "step": 843
+ },
+ {
+ "epoch": 21.27,
+ "learning_rate": 0.00016185897435897436,
+ "loss": 0.1234,
+ "step": 844
+ },
+ {
+ "epoch": 21.29,
+ "learning_rate": 0.00016145833333333336,
+ "loss": 0.1226,
+ "step": 845
+ },
+ {
+ "epoch": 21.32,
+ "learning_rate": 0.00016105769230769233,
+ "loss": 0.1224,
+ "step": 846
+ },
+ {
+ "epoch": 21.34,
+ "learning_rate": 0.0001606570512820513,
+ "loss": 0.1266,
+ "step": 847
+ },
+ {
+ "epoch": 21.37,
+ "learning_rate": 0.00016025641025641028,
+ "loss": 0.1267,
+ "step": 848
+ },
+ {
+ "epoch": 21.39,
+ "learning_rate": 0.00015985576923076923,
+ "loss": 0.129,
+ "step": 849
+ },
+ {
+ "epoch": 21.42,
+ "learning_rate": 0.0001594551282051282,
+ "loss": 0.1213,
+ "step": 850
+ },
+ {
+ "epoch": 21.44,
+ "learning_rate": 0.00015905448717948718,
+ "loss": 0.1273,
+ "step": 851
+ },
+ {
+ "epoch": 21.47,
+ "learning_rate": 0.00015865384615384616,
+ "loss": 0.1352,
+ "step": 852
+ },
+ {
+ "epoch": 21.49,
+ "learning_rate": 0.00015825320512820513,
+ "loss": 0.1328,
+ "step": 853
+ },
+ {
+ "epoch": 21.52,
+ "learning_rate": 0.0001578525641025641,
+ "loss": 0.1253,
+ "step": 854
+ },
+ {
+ "epoch": 21.54,
+ "learning_rate": 0.00015745192307692308,
+ "loss": 0.1267,
+ "step": 855
+ },
+ {
+ "epoch": 21.57,
+ "learning_rate": 0.00015705128205128205,
+ "loss": 0.1325,
+ "step": 856
+ },
+ {
+ "epoch": 21.59,
+ "learning_rate": 0.00015665064102564103,
+ "loss": 0.1267,
+ "step": 857
+ },
+ {
+ "epoch": 21.62,
+ "learning_rate": 0.00015625,
+ "loss": 0.1326,
+ "step": 858
+ },
+ {
+ "epoch": 21.64,
+ "learning_rate": 0.00015584935897435898,
+ "loss": 0.1319,
+ "step": 859
+ },
+ {
+ "epoch": 21.67,
+ "learning_rate": 0.00015544871794871795,
+ "loss": 0.1262,
+ "step": 860
+ },
+ {
+ "epoch": 21.69,
+ "learning_rate": 0.00015504807692307693,
+ "loss": 0.1334,
+ "step": 861
+ },
+ {
+ "epoch": 21.72,
+ "learning_rate": 0.0001546474358974359,
+ "loss": 0.1347,
+ "step": 862
+ },
+ {
+ "epoch": 21.74,
+ "learning_rate": 0.00015424679487179488,
+ "loss": 0.1322,
+ "step": 863
+ },
+ {
+ "epoch": 21.77,
+ "learning_rate": 0.00015384615384615385,
+ "loss": 0.1321,
+ "step": 864
+ },
+ {
+ "epoch": 21.8,
+ "learning_rate": 0.00015344551282051283,
+ "loss": 0.128,
+ "step": 865
+ },
+ {
+ "epoch": 21.82,
+ "learning_rate": 0.0001530448717948718,
+ "loss": 0.1365,
+ "step": 866
+ },
+ {
+ "epoch": 21.85,
+ "learning_rate": 0.00015264423076923078,
+ "loss": 0.137,
+ "step": 867
+ },
+ {
+ "epoch": 21.87,
+ "learning_rate": 0.00015224358974358972,
+ "loss": 0.1322,
+ "step": 868
+ },
+ {
+ "epoch": 21.9,
+ "learning_rate": 0.0001518429487179487,
+ "loss": 0.1341,
+ "step": 869
+ },
+ {
+ "epoch": 21.92,
+ "learning_rate": 0.00015144230769230767,
+ "loss": 0.1334,
+ "step": 870
+ },
+ {
+ "epoch": 21.95,
+ "learning_rate": 0.00015104166666666665,
+ "loss": 0.1339,
+ "step": 871
+ },
+ {
+ "epoch": 21.97,
+ "learning_rate": 0.00015064102564102565,
+ "loss": 0.1301,
+ "step": 872
+ },
+ {
+ "epoch": 22.0,
+ "learning_rate": 0.00015024038461538462,
+ "loss": 0.1325,
+ "step": 873
+ },
+ {
+ "epoch": 22.02,
+ "learning_rate": 0.0001498397435897436,
+ "loss": 0.1153,
+ "step": 874
+ },
+ {
+ "epoch": 22.05,
+ "learning_rate": 0.00014943910256410257,
+ "loss": 0.1113,
+ "step": 875
+ },
+ {
+ "epoch": 22.07,
+ "learning_rate": 0.00014903846153846155,
+ "loss": 0.1138,
+ "step": 876
+ },
+ {
+ "epoch": 22.1,
+ "learning_rate": 0.00014863782051282052,
+ "loss": 0.1136,
+ "step": 877
+ },
+ {
+ "epoch": 22.12,
+ "learning_rate": 0.0001482371794871795,
+ "loss": 0.1096,
+ "step": 878
+ },
+ {
+ "epoch": 22.15,
+ "learning_rate": 0.00014783653846153847,
+ "loss": 0.116,
+ "step": 879
+ },
+ {
+ "epoch": 22.17,
+ "learning_rate": 0.00014743589743589745,
+ "loss": 0.1122,
+ "step": 880
+ },
+ {
+ "epoch": 22.2,
+ "learning_rate": 0.00014703525641025642,
+ "loss": 0.1132,
+ "step": 881
+ },
+ {
+ "epoch": 22.22,
+ "learning_rate": 0.0001466346153846154,
+ "loss": 0.1128,
+ "step": 882
+ },
+ {
+ "epoch": 22.25,
+ "learning_rate": 0.00014623397435897437,
+ "loss": 0.1134,
+ "step": 883
+ },
+ {
+ "epoch": 22.27,
+ "learning_rate": 0.00014583333333333335,
+ "loss": 0.1065,
+ "step": 884
+ },
+ {
+ "epoch": 22.3,
+ "learning_rate": 0.00014543269230769232,
+ "loss": 0.117,
+ "step": 885
+ },
+ {
+ "epoch": 22.32,
+ "learning_rate": 0.0001450320512820513,
+ "loss": 0.112,
+ "step": 886
+ },
+ {
+ "epoch": 22.35,
+ "learning_rate": 0.00014463141025641027,
+ "loss": 0.1145,
+ "step": 887
+ },
+ {
+ "epoch": 22.37,
+ "learning_rate": 0.00014423076923076922,
+ "loss": 0.1128,
+ "step": 888
+ },
+ {
+ "epoch": 22.4,
+ "learning_rate": 0.0001438301282051282,
+ "loss": 0.1132,
+ "step": 889
+ },
+ {
+ "epoch": 22.43,
+ "learning_rate": 0.00014342948717948717,
+ "loss": 0.1143,
+ "step": 890
+ },
+ {
+ "epoch": 22.45,
+ "learning_rate": 0.00014302884615384614,
+ "loss": 0.1172,
+ "step": 891
+ },
+ {
+ "epoch": 22.48,
+ "learning_rate": 0.00014262820512820512,
+ "loss": 0.1143,
+ "step": 892
+ },
+ {
+ "epoch": 22.5,
+ "learning_rate": 0.0001422275641025641,
+ "loss": 0.1164,
+ "step": 893
+ },
+ {
+ "epoch": 22.53,
+ "learning_rate": 0.00014182692307692307,
+ "loss": 0.1197,
+ "step": 894
+ },
+ {
+ "epoch": 22.55,
+ "learning_rate": 0.00014142628205128204,
+ "loss": 0.1152,
+ "step": 895
+ },
+ {
+ "epoch": 22.58,
+ "learning_rate": 0.00014102564102564101,
+ "loss": 0.1243,
+ "step": 896
+ },
+ {
+ "epoch": 22.6,
+ "learning_rate": 0.00014062500000000002,
+ "loss": 0.1167,
+ "step": 897
+ },
+ {
+ "epoch": 22.63,
+ "learning_rate": 0.000140224358974359,
+ "loss": 0.1224,
+ "step": 898
+ },
+ {
+ "epoch": 22.65,
+ "learning_rate": 0.00013982371794871797,
+ "loss": 0.1192,
+ "step": 899
+ },
+ {
+ "epoch": 22.68,
+ "learning_rate": 0.00013942307692307694,
+ "loss": 0.1184,
+ "step": 900
+ },
+ {
+ "epoch": 22.7,
+ "learning_rate": 0.00013902243589743592,
+ "loss": 0.1169,
+ "step": 901
+ },
+ {
+ "epoch": 22.73,
+ "learning_rate": 0.0001386217948717949,
+ "loss": 0.1169,
+ "step": 902
+ },
+ {
+ "epoch": 22.75,
+ "learning_rate": 0.00013822115384615386,
+ "loss": 0.1177,
+ "step": 903
+ },
+ {
+ "epoch": 22.78,
+ "learning_rate": 0.00013782051282051284,
+ "loss": 0.1175,
+ "step": 904
+ },
+ {
+ "epoch": 22.8,
+ "learning_rate": 0.00013741987179487181,
+ "loss": 0.1164,
+ "step": 905
+ },
+ {
+ "epoch": 22.83,
+ "learning_rate": 0.0001370192307692308,
+ "loss": 0.1215,
+ "step": 906
+ },
+ {
+ "epoch": 22.85,
+ "learning_rate": 0.00013661858974358974,
+ "loss": 0.1204,
+ "step": 907
+ },
+ {
+ "epoch": 22.88,
+ "learning_rate": 0.0001362179487179487,
+ "loss": 0.1208,
+ "step": 908
+ },
+ {
+ "epoch": 22.9,
+ "learning_rate": 0.00013581730769230769,
+ "loss": 0.1175,
+ "step": 909
+ },
+ {
+ "epoch": 22.93,
+ "learning_rate": 0.00013541666666666666,
+ "loss": 0.1214,
+ "step": 910
+ },
+ {
+ "epoch": 22.95,
+ "learning_rate": 0.00013501602564102564,
+ "loss": 0.1243,
+ "step": 911
+ },
+ {
+ "epoch": 22.98,
+ "learning_rate": 0.0001346153846153846,
+ "loss": 0.1266,
+ "step": 912
+ },
+ {
+ "epoch": 23.0,
+ "learning_rate": 0.00013421474358974358,
+ "loss": 0.1226,
+ "step": 913
+ },
+ {
+ "epoch": 23.03,
+ "learning_rate": 0.00013381410256410256,
+ "loss": 0.1058,
+ "step": 914
+ },
+ {
+ "epoch": 23.06,
+ "learning_rate": 0.00013341346153846153,
+ "loss": 0.1034,
+ "step": 915
+ },
+ {
+ "epoch": 23.08,
+ "learning_rate": 0.0001330128205128205,
+ "loss": 0.1026,
+ "step": 916
+ },
+ {
+ "epoch": 23.11,
+ "learning_rate": 0.00013261217948717948,
+ "loss": 0.1026,
+ "step": 917
+ },
+ {
+ "epoch": 23.13,
+ "learning_rate": 0.00013221153846153846,
+ "loss": 0.1028,
+ "step": 918
+ },
+ {
+ "epoch": 23.16,
+ "learning_rate": 0.00013181089743589743,
+ "loss": 0.1024,
+ "step": 919
+ },
+ {
+ "epoch": 23.18,
+ "learning_rate": 0.0001314102564102564,
+ "loss": 0.1042,
+ "step": 920
+ },
+ {
+ "epoch": 23.21,
+ "learning_rate": 0.00013100961538461538,
+ "loss": 0.0992,
+ "step": 921
+ },
+ {
+ "epoch": 23.23,
+ "learning_rate": 0.00013060897435897436,
+ "loss": 0.1048,
+ "step": 922
+ },
+ {
+ "epoch": 23.26,
+ "learning_rate": 0.00013020833333333336,
+ "loss": 0.1006,
+ "step": 923
+ },
+ {
+ "epoch": 23.28,
+ "learning_rate": 0.00012980769230769233,
+ "loss": 0.1041,
+ "step": 924
+ },
+ {
+ "epoch": 23.31,
+ "learning_rate": 0.0001294070512820513,
+ "loss": 0.1036,
+ "step": 925
+ },
+ {
+ "epoch": 23.33,
+ "learning_rate": 0.00012900641025641028,
+ "loss": 0.1045,
+ "step": 926
+ },
+ {
+ "epoch": 23.36,
+ "learning_rate": 0.00012860576923076923,
+ "loss": 0.1048,
+ "step": 927
+ },
+ {
+ "epoch": 23.38,
+ "learning_rate": 0.0001282051282051282,
+ "loss": 0.1082,
+ "step": 928
+ },
+ {
+ "epoch": 23.41,
+ "learning_rate": 0.00012780448717948718,
+ "loss": 0.1092,
+ "step": 929
+ },
+ {
+ "epoch": 23.43,
+ "learning_rate": 0.00012740384615384615,
+ "loss": 0.1085,
+ "step": 930
+ },
+ {
+ "epoch": 23.46,
+ "learning_rate": 0.00012700320512820513,
+ "loss": 0.1037,
+ "step": 931
+ },
+ {
+ "epoch": 23.48,
+ "learning_rate": 0.0001266025641025641,
+ "loss": 0.109,
+ "step": 932
+ },
+ {
+ "epoch": 23.51,
+ "learning_rate": 0.00012620192307692308,
+ "loss": 0.1076,
+ "step": 933
+ },
+ {
+ "epoch": 23.53,
+ "learning_rate": 0.00012580128205128205,
+ "loss": 0.1062,
+ "step": 934
+ },
+ {
+ "epoch": 23.56,
+ "learning_rate": 0.00012540064102564103,
+ "loss": 0.1049,
+ "step": 935
+ },
+ {
+ "epoch": 23.58,
+ "learning_rate": 0.000125,
+ "loss": 0.1057,
+ "step": 936
+ },
+ {
+ "epoch": 23.61,
+ "learning_rate": 0.00012459935897435898,
+ "loss": 0.1106,
+ "step": 937
+ },
+ {
+ "epoch": 23.63,
+ "learning_rate": 0.00012419871794871795,
+ "loss": 0.1134,
+ "step": 938
+ },
+ {
+ "epoch": 23.66,
+ "learning_rate": 0.00012379807692307693,
+ "loss": 0.1124,
+ "step": 939
+ },
+ {
+ "epoch": 23.69,
+ "learning_rate": 0.0001233974358974359,
+ "loss": 0.105,
+ "step": 940
+ },
+ {
+ "epoch": 23.71,
+ "learning_rate": 0.00012299679487179488,
+ "loss": 0.113,
+ "step": 941
+ },
+ {
+ "epoch": 23.74,
+ "learning_rate": 0.00012259615384615385,
+ "loss": 0.1104,
+ "step": 942
+ },
+ {
+ "epoch": 23.76,
+ "learning_rate": 0.00012219551282051283,
+ "loss": 0.1136,
+ "step": 943
+ },
+ {
+ "epoch": 23.79,
+ "learning_rate": 0.0001217948717948718,
+ "loss": 0.1104,
+ "step": 944
+ },
+ {
+ "epoch": 23.81,
+ "learning_rate": 0.00012139423076923077,
+ "loss": 0.1054,
+ "step": 945
+ },
+ {
+ "epoch": 23.84,
+ "learning_rate": 0.00012099358974358975,
+ "loss": 0.1109,
+ "step": 946
+ },
+ {
+ "epoch": 23.86,
+ "learning_rate": 0.00012059294871794872,
+ "loss": 0.1062,
+ "step": 947
+ },
+ {
+ "epoch": 23.89,
+ "learning_rate": 0.0001201923076923077,
+ "loss": 0.1185,
+ "step": 948
+ },
+ {
+ "epoch": 23.91,
+ "learning_rate": 0.00011979166666666667,
+ "loss": 0.1138,
+ "step": 949
+ },
+ {
+ "epoch": 23.94,
+ "learning_rate": 0.00011939102564102565,
+ "loss": 0.1119,
+ "step": 950
+ },
+ {
+ "epoch": 23.96,
+ "learning_rate": 0.00011899038461538461,
+ "loss": 0.1114,
+ "step": 951
+ },
+ {
+ "epoch": 23.99,
+ "learning_rate": 0.00011858974358974358,
+ "loss": 0.1122,
+ "step": 952
+ },
+ {
+ "epoch": 24.01,
+ "learning_rate": 0.00011818910256410256,
+ "loss": 0.1051,
+ "step": 953
+ },
+ {
+ "epoch": 24.04,
+ "learning_rate": 0.00011778846153846153,
+ "loss": 0.0924,
+ "step": 954
+ },
+ {
+ "epoch": 24.06,
+ "learning_rate": 0.00011738782051282051,
+ "loss": 0.0977,
+ "step": 955
+ },
+ {
+ "epoch": 24.09,
+ "learning_rate": 0.0001169871794871795,
+ "loss": 0.0945,
+ "step": 956
+ },
+ {
+ "epoch": 24.11,
+ "learning_rate": 0.00011658653846153847,
+ "loss": 0.0958,
+ "step": 957
+ },
+ {
+ "epoch": 24.14,
+ "learning_rate": 0.00011618589743589745,
+ "loss": 0.0934,
+ "step": 958
+ },
+ {
+ "epoch": 24.16,
+ "learning_rate": 0.00011578525641025642,
+ "loss": 0.0952,
+ "step": 959
+ },
+ {
+ "epoch": 24.19,
+ "learning_rate": 0.0001153846153846154,
+ "loss": 0.0993,
+ "step": 960
+ },
+ {
+ "epoch": 24.21,
+ "learning_rate": 0.00011498397435897436,
+ "loss": 0.0978,
+ "step": 961
+ },
+ {
+ "epoch": 24.24,
+ "learning_rate": 0.00011458333333333333,
+ "loss": 0.0984,
+ "step": 962
+ },
+ {
+ "epoch": 24.26,
+ "learning_rate": 0.0001141826923076923,
+ "loss": 0.1019,
+ "step": 963
+ },
+ {
+ "epoch": 24.29,
+ "learning_rate": 0.00011378205128205128,
+ "loss": 0.0994,
+ "step": 964
+ },
+ {
+ "epoch": 24.31,
+ "learning_rate": 0.00011338141025641026,
+ "loss": 0.1006,
+ "step": 965
+ },
+ {
+ "epoch": 24.34,
+ "learning_rate": 0.00011298076923076923,
+ "loss": 0.0971,
+ "step": 966
+ },
+ {
+ "epoch": 24.37,
+ "learning_rate": 0.0001125801282051282,
+ "loss": 0.1008,
+ "step": 967
+ },
+ {
+ "epoch": 24.39,
+ "learning_rate": 0.00011217948717948718,
+ "loss": 0.0948,
+ "step": 968
+ },
+ {
+ "epoch": 24.42,
+ "learning_rate": 0.00011177884615384617,
+ "loss": 0.1026,
+ "step": 969
+ },
+ {
+ "epoch": 24.44,
+ "learning_rate": 0.00011137820512820514,
+ "loss": 0.0966,
+ "step": 970
+ },
+ {
+ "epoch": 24.47,
+ "learning_rate": 0.0001109775641025641,
+ "loss": 0.1004,
+ "step": 971
+ },
+ {
+ "epoch": 24.49,
+ "learning_rate": 0.00011057692307692308,
+ "loss": 0.1031,
+ "step": 972
+ },
+ {
+ "epoch": 24.52,
+ "learning_rate": 0.00011017628205128205,
+ "loss": 0.0984,
+ "step": 973
+ },
+ {
+ "epoch": 24.54,
+ "learning_rate": 0.00010977564102564103,
+ "loss": 0.0988,
+ "step": 974
+ },
+ {
+ "epoch": 24.57,
+ "learning_rate": 0.000109375,
+ "loss": 0.1048,
+ "step": 975
+ },
+ {
+ "epoch": 24.59,
+ "learning_rate": 0.00010897435897435898,
+ "loss": 0.0987,
+ "step": 976
+ },
+ {
+ "epoch": 24.62,
+ "learning_rate": 0.00010857371794871795,
+ "loss": 0.1026,
+ "step": 977
+ },
+ {
+ "epoch": 24.64,
+ "learning_rate": 0.00010817307692307693,
+ "loss": 0.1015,
+ "step": 978
+ },
+ {
+ "epoch": 24.67,
+ "learning_rate": 0.0001077724358974359,
+ "loss": 0.102,
+ "step": 979
+ },
+ {
+ "epoch": 24.69,
+ "learning_rate": 0.00010737179487179486,
+ "loss": 0.1005,
+ "step": 980
+ },
+ {
+ "epoch": 24.72,
+ "learning_rate": 0.00010697115384615384,
+ "loss": 0.1042,
+ "step": 981
+ },
+ {
+ "epoch": 24.74,
+ "learning_rate": 0.00010657051282051283,
+ "loss": 0.1046,
+ "step": 982
+ },
+ {
+ "epoch": 24.77,
+ "learning_rate": 0.0001061698717948718,
+ "loss": 0.0976,
+ "step": 983
+ },
+ {
+ "epoch": 24.79,
+ "learning_rate": 0.00010576923076923077,
+ "loss": 0.1014,
+ "step": 984
+ },
+ {
+ "epoch": 24.82,
+ "learning_rate": 0.00010536858974358975,
+ "loss": 0.1028,
+ "step": 985
+ },
+ {
+ "epoch": 24.84,
+ "learning_rate": 0.00010496794871794872,
+ "loss": 0.1022,
+ "step": 986
+ },
+ {
+ "epoch": 24.87,
+ "learning_rate": 0.0001045673076923077,
+ "loss": 0.1021,
+ "step": 987
+ },
+ {
+ "epoch": 24.89,
+ "learning_rate": 0.00010416666666666667,
+ "loss": 0.1047,
+ "step": 988
+ },
+ {
+ "epoch": 24.92,
+ "learning_rate": 0.00010376602564102565,
+ "loss": 0.1062,
+ "step": 989
+ },
+ {
+ "epoch": 24.94,
+ "learning_rate": 0.00010336538461538461,
+ "loss": 0.1028,
+ "step": 990
+ },
+ {
+ "epoch": 24.97,
+ "learning_rate": 0.00010296474358974358,
+ "loss": 0.1051,
+ "step": 991
+ },
+ {
+ "epoch": 25.0,
+ "learning_rate": 0.00010256410256410256,
+ "loss": 0.1047,
+ "step": 992
+ },
+ {
+ "epoch": 25.02,
+ "learning_rate": 0.00010216346153846153,
+ "loss": 0.097,
+ "step": 993
+ },
+ {
+ "epoch": 25.05,
+ "learning_rate": 0.00010176282051282051,
+ "loss": 0.0963,
+ "step": 994
+ },
+ {
+ "epoch": 25.07,
+ "learning_rate": 0.0001013621794871795,
+ "loss": 0.0888,
+ "step": 995
+ },
+ {
+ "epoch": 25.1,
+ "learning_rate": 0.00010096153846153847,
+ "loss": 0.089,
+ "step": 996
+ },
+ {
+ "epoch": 25.12,
+ "learning_rate": 0.00010056089743589745,
+ "loss": 0.0911,
+ "step": 997
+ },
+ {
+ "epoch": 25.15,
+ "learning_rate": 0.00010016025641025642,
+ "loss": 0.0916,
+ "step": 998
+ },
+ {
+ "epoch": 25.17,
+ "learning_rate": 9.97596153846154e-05,
+ "loss": 0.089,
+ "step": 999
+ },
+ {
+ "epoch": 25.2,
+ "learning_rate": 9.935897435897436e-05,
+ "loss": 0.0883,
+ "step": 1000
+ },
+ {
+ "epoch": 25.22,
+ "learning_rate": 9.895833333333333e-05,
+ "loss": 0.0999,
+ "step": 1001
+ },
+ {
+ "epoch": 25.25,
+ "learning_rate": 9.85576923076923e-05,
+ "loss": 0.0918,
+ "step": 1002
+ },
+ {
+ "epoch": 25.27,
+ "learning_rate": 9.815705128205128e-05,
+ "loss": 0.0953,
+ "step": 1003
+ },
+ {
+ "epoch": 25.3,
+ "learning_rate": 9.775641025641025e-05,
+ "loss": 0.0917,
+ "step": 1004
+ },
+ {
+ "epoch": 25.32,
+ "learning_rate": 9.735576923076923e-05,
+ "loss": 0.0913,
+ "step": 1005
+ },
+ {
+ "epoch": 25.35,
+ "learning_rate": 9.69551282051282e-05,
+ "loss": 0.0912,
+ "step": 1006
+ },
+ {
+ "epoch": 25.37,
+ "learning_rate": 9.655448717948718e-05,
+ "loss": 0.0897,
+ "step": 1007
+ },
+ {
+ "epoch": 25.4,
+ "learning_rate": 9.615384615384617e-05,
+ "loss": 0.0924,
+ "step": 1008
+ },
+ {
+ "epoch": 25.42,
+ "learning_rate": 9.575320512820514e-05,
+ "loss": 0.0939,
+ "step": 1009
+ },
+ {
+ "epoch": 25.45,
+ "learning_rate": 9.53525641025641e-05,
+ "loss": 0.0931,
+ "step": 1010
+ },
+ {
+ "epoch": 25.47,
+ "learning_rate": 9.495192307692308e-05,
+ "loss": 0.0919,
+ "step": 1011
+ },
+ {
+ "epoch": 25.5,
+ "learning_rate": 9.455128205128205e-05,
+ "loss": 0.0926,
+ "step": 1012
+ },
+ {
+ "epoch": 25.52,
+ "learning_rate": 9.415064102564103e-05,
+ "loss": 0.0958,
+ "step": 1013
+ },
+ {
+ "epoch": 25.55,
+ "learning_rate": 9.375e-05,
+ "loss": 0.097,
+ "step": 1014
+ },
+ {
+ "epoch": 25.57,
+ "learning_rate": 9.334935897435898e-05,
+ "loss": 0.0946,
+ "step": 1015
+ },
+ {
+ "epoch": 25.6,
+ "learning_rate": 9.294871794871795e-05,
+ "loss": 0.0943,
+ "step": 1016
+ },
+ {
+ "epoch": 25.63,
+ "learning_rate": 9.254807692307693e-05,
+ "loss": 0.0931,
+ "step": 1017
+ },
+ {
+ "epoch": 25.65,
+ "learning_rate": 9.21474358974359e-05,
+ "loss": 0.094,
+ "step": 1018
+ },
+ {
+ "epoch": 25.68,
+ "learning_rate": 9.174679487179486e-05,
+ "loss": 0.0971,
+ "step": 1019
+ },
+ {
+ "epoch": 25.7,
+ "learning_rate": 9.134615384615384e-05,
+ "loss": 0.0972,
+ "step": 1020
+ },
+ {
+ "epoch": 25.73,
+ "learning_rate": 9.094551282051282e-05,
+ "loss": 0.1061,
+ "step": 1021
+ },
+ {
+ "epoch": 25.75,
+ "learning_rate": 9.05448717948718e-05,
+ "loss": 0.0974,
+ "step": 1022
+ },
+ {
+ "epoch": 25.78,
+ "learning_rate": 9.014423076923077e-05,
+ "loss": 0.0964,
+ "step": 1023
+ },
+ {
+ "epoch": 25.8,
+ "learning_rate": 8.974358974358975e-05,
+ "loss": 0.095,
+ "step": 1024
+ },
+ {
+ "epoch": 25.83,
+ "learning_rate": 8.934294871794872e-05,
+ "loss": 0.1016,
+ "step": 1025
+ },
+ {
+ "epoch": 25.85,
+ "learning_rate": 8.89423076923077e-05,
+ "loss": 0.0959,
+ "step": 1026
+ },
+ {
+ "epoch": 25.88,
+ "learning_rate": 8.854166666666667e-05,
+ "loss": 0.1008,
+ "step": 1027
+ },
+ {
+ "epoch": 25.9,
+ "learning_rate": 8.814102564102565e-05,
+ "loss": 0.1004,
+ "step": 1028
+ },
+ {
+ "epoch": 25.93,
+ "learning_rate": 8.774038461538461e-05,
+ "loss": 0.0959,
+ "step": 1029
+ },
+ {
+ "epoch": 25.95,
+ "learning_rate": 8.733974358974358e-05,
+ "loss": 0.0958,
+ "step": 1030
+ },
+ {
+ "epoch": 25.98,
+ "learning_rate": 8.693910256410256e-05,
+ "loss": 0.0961,
+ "step": 1031
+ },
+ {
+ "epoch": 26.0,
+ "learning_rate": 8.653846153846153e-05,
+ "loss": 0.0957,
+ "step": 1032
+ },
+ {
+ "epoch": 26.03,
+ "learning_rate": 8.613782051282051e-05,
+ "loss": 0.0855,
+ "step": 1033
+ },
+ {
+ "epoch": 26.05,
+ "learning_rate": 8.57371794871795e-05,
+ "loss": 0.0835,
+ "step": 1034
+ },
+ {
+ "epoch": 26.08,
+ "learning_rate": 8.533653846153847e-05,
+ "loss": 0.0886,
+ "step": 1035
+ },
+ {
+ "epoch": 26.1,
+ "learning_rate": 8.493589743589745e-05,
+ "loss": 0.0918,
+ "step": 1036
+ },
+ {
+ "epoch": 26.13,
+ "learning_rate": 8.453525641025642e-05,
+ "loss": 0.0837,
+ "step": 1037
+ },
+ {
+ "epoch": 26.15,
+ "learning_rate": 8.41346153846154e-05,
+ "loss": 0.088,
+ "step": 1038
+ },
+ {
+ "epoch": 26.18,
+ "learning_rate": 8.373397435897436e-05,
+ "loss": 0.0928,
+ "step": 1039
+ },
+ {
+ "epoch": 26.2,
+ "learning_rate": 8.333333333333333e-05,
+ "loss": 0.0871,
+ "step": 1040
+ },
+ {
+ "epoch": 26.23,
+ "learning_rate": 8.29326923076923e-05,
+ "loss": 0.0863,
+ "step": 1041
+ },
+ {
+ "epoch": 26.26,
+ "learning_rate": 8.253205128205128e-05,
+ "loss": 0.0876,
+ "step": 1042
+ },
+ {
+ "epoch": 26.28,
+ "learning_rate": 8.213141025641025e-05,
+ "loss": 0.0888,
+ "step": 1043
+ },
+ {
+ "epoch": 26.31,
+ "learning_rate": 8.173076923076923e-05,
+ "loss": 0.0898,
+ "step": 1044
+ },
+ {
+ "epoch": 26.33,
+ "learning_rate": 8.13301282051282e-05,
+ "loss": 0.09,
+ "step": 1045
+ },
+ {
+ "epoch": 26.36,
+ "learning_rate": 8.092948717948718e-05,
+ "loss": 0.0906,
+ "step": 1046
+ },
+ {
+ "epoch": 26.38,
+ "learning_rate": 8.052884615384617e-05,
+ "loss": 0.091,
+ "step": 1047
+ },
+ {
+ "epoch": 26.41,
+ "learning_rate": 8.012820512820514e-05,
+ "loss": 0.0885,
+ "step": 1048
+ },
+ {
+ "epoch": 26.43,
+ "learning_rate": 7.97275641025641e-05,
+ "loss": 0.0896,
+ "step": 1049
+ },
+ {
+ "epoch": 26.46,
+ "learning_rate": 7.932692307692308e-05,
+ "loss": 0.0891,
+ "step": 1050
+ },
+ {
+ "epoch": 26.48,
+ "learning_rate": 7.892628205128205e-05,
+ "loss": 0.0875,
+ "step": 1051
+ },
+ {
+ "epoch": 26.51,
+ "learning_rate": 7.852564102564103e-05,
+ "loss": 0.0906,
+ "step": 1052
+ },
+ {
+ "epoch": 26.53,
+ "learning_rate": 7.8125e-05,
+ "loss": 0.0913,
+ "step": 1053
+ },
+ {
+ "epoch": 26.56,
+ "learning_rate": 7.772435897435898e-05,
+ "loss": 0.0898,
+ "step": 1054
+ },
+ {
+ "epoch": 26.58,
+ "learning_rate": 7.732371794871795e-05,
+ "loss": 0.0887,
+ "step": 1055
+ },
+ {
+ "epoch": 26.61,
+ "learning_rate": 7.692307692307693e-05,
+ "loss": 0.0916,
+ "step": 1056
+ },
+ {
+ "epoch": 26.63,
+ "learning_rate": 7.65224358974359e-05,
+ "loss": 0.0934,
+ "step": 1057
+ },
+ {
+ "epoch": 26.66,
+ "learning_rate": 7.612179487179486e-05,
+ "loss": 0.0864,
+ "step": 1058
+ },
+ {
+ "epoch": 26.68,
+ "learning_rate": 7.572115384615384e-05,
+ "loss": 0.0929,
+ "step": 1059
+ },
+ {
+ "epoch": 26.71,
+ "learning_rate": 7.532051282051282e-05,
+ "loss": 0.0893,
+ "step": 1060
+ },
+ {
+ "epoch": 26.73,
+ "learning_rate": 7.49198717948718e-05,
+ "loss": 0.0926,
+ "step": 1061
+ },
+ {
+ "epoch": 26.76,
+ "learning_rate": 7.451923076923077e-05,
+ "loss": 0.0931,
+ "step": 1062
+ },
+ {
+ "epoch": 26.78,
+ "learning_rate": 7.411858974358975e-05,
+ "loss": 0.0921,
+ "step": 1063
+ },
+ {
+ "epoch": 26.81,
+ "learning_rate": 7.371794871794872e-05,
+ "loss": 0.0882,
+ "step": 1064
+ },
+ {
+ "epoch": 26.83,
+ "learning_rate": 7.33173076923077e-05,
+ "loss": 0.0937,
+ "step": 1065
+ },
+ {
+ "epoch": 26.86,
+ "learning_rate": 7.291666666666667e-05,
+ "loss": 0.0913,
+ "step": 1066
+ },
+ {
+ "epoch": 26.89,
+ "learning_rate": 7.251602564102565e-05,
+ "loss": 0.0925,
+ "step": 1067
+ },
+ {
+ "epoch": 26.91,
+ "learning_rate": 7.211538461538461e-05,
+ "loss": 0.0942,
+ "step": 1068
+ },
+ {
+ "epoch": 26.94,
+ "learning_rate": 7.171474358974358e-05,
+ "loss": 0.097,
+ "step": 1069
+ },
+ {
+ "epoch": 26.96,
+ "learning_rate": 7.131410256410256e-05,
+ "loss": 0.0922,
+ "step": 1070
+ },
+ {
+ "epoch": 26.99,
+ "learning_rate": 7.091346153846153e-05,
+ "loss": 0.0925,
+ "step": 1071
+ },
+ {
+ "epoch": 27.01,
+ "learning_rate": 7.051282051282051e-05,
+ "loss": 0.0887,
+ "step": 1072
+ },
+ {
+ "epoch": 27.04,
+ "learning_rate": 7.01121794871795e-05,
+ "loss": 0.0831,
+ "step": 1073
+ },
+ {
+ "epoch": 27.06,
+ "learning_rate": 6.971153846153847e-05,
+ "loss": 0.084,
+ "step": 1074
+ },
+ {
+ "epoch": 27.09,
+ "learning_rate": 6.931089743589744e-05,
+ "loss": 0.083,
+ "step": 1075
+ },
+ {
+ "epoch": 27.11,
+ "learning_rate": 6.891025641025642e-05,
+ "loss": 0.0857,
+ "step": 1076
+ },
+ {
+ "epoch": 27.14,
+ "learning_rate": 6.85096153846154e-05,
+ "loss": 0.0837,
+ "step": 1077
+ },
+ {
+ "epoch": 27.16,
+ "learning_rate": 6.810897435897436e-05,
+ "loss": 0.0854,
+ "step": 1078
+ },
+ {
+ "epoch": 27.19,
+ "learning_rate": 6.770833333333333e-05,
+ "loss": 0.0877,
+ "step": 1079
+ },
+ {
+ "epoch": 27.21,
+ "learning_rate": 6.73076923076923e-05,
+ "loss": 0.0861,
+ "step": 1080
+ },
+ {
+ "epoch": 27.24,
+ "learning_rate": 6.690705128205128e-05,
+ "loss": 0.0825,
+ "step": 1081
+ },
+ {
+ "epoch": 27.26,
+ "learning_rate": 6.650641025641025e-05,
+ "loss": 0.0831,
+ "step": 1082
+ },
+ {
+ "epoch": 27.29,
+ "learning_rate": 6.610576923076923e-05,
+ "loss": 0.086,
+ "step": 1083
+ },
+ {
+ "epoch": 27.31,
+ "learning_rate": 6.57051282051282e-05,
+ "loss": 0.0838,
+ "step": 1084
+ },
+ {
+ "epoch": 27.34,
+ "learning_rate": 6.530448717948718e-05,
+ "loss": 0.0857,
+ "step": 1085
+ },
+ {
+ "epoch": 27.36,
+ "learning_rate": 6.490384615384617e-05,
+ "loss": 0.0836,
+ "step": 1086
+ },
+ {
+ "epoch": 27.39,
+ "learning_rate": 6.450320512820514e-05,
+ "loss": 0.0875,
+ "step": 1087
+ },
+ {
+ "epoch": 27.41,
+ "learning_rate": 6.41025641025641e-05,
+ "loss": 0.0848,
+ "step": 1088
+ },
+ {
+ "epoch": 27.44,
+ "learning_rate": 6.370192307692308e-05,
+ "loss": 0.084,
+ "step": 1089
+ },
+ {
+ "epoch": 27.46,
+ "learning_rate": 6.330128205128205e-05,
+ "loss": 0.0845,
+ "step": 1090
+ },
+ {
+ "epoch": 27.49,
+ "learning_rate": 6.290064102564103e-05,
+ "loss": 0.0876,
+ "step": 1091
+ },
+ {
+ "epoch": 27.51,
+ "learning_rate": 6.25e-05,
+ "loss": 0.0907,
+ "step": 1092
+ },
+ {
+ "epoch": 27.54,
+ "learning_rate": 6.209935897435898e-05,
+ "loss": 0.0864,
+ "step": 1093
+ },
+ {
+ "epoch": 27.57,
+ "learning_rate": 6.169871794871795e-05,
+ "loss": 0.087,
+ "step": 1094
+ },
+ {
+ "epoch": 27.59,
+ "learning_rate": 6.129807692307693e-05,
+ "loss": 0.0902,
+ "step": 1095
+ },
+ {
+ "epoch": 27.62,
+ "learning_rate": 6.08974358974359e-05,
+ "loss": 0.0862,
+ "step": 1096
+ },
+ {
+ "epoch": 27.64,
+ "learning_rate": 6.0496794871794875e-05,
+ "loss": 0.0864,
+ "step": 1097
+ },
+ {
+ "epoch": 27.67,
+ "learning_rate": 6.009615384615385e-05,
+ "loss": 0.0868,
+ "step": 1098
+ },
+ {
+ "epoch": 27.69,
+ "learning_rate": 5.9695512820512824e-05,
+ "loss": 0.0905,
+ "step": 1099
+ },
+ {
+ "epoch": 27.72,
+ "learning_rate": 5.929487179487179e-05,
+ "loss": 0.0881,
+ "step": 1100
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 6.19395154565591e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-1100/training_args.bin b/checkpoint-1100/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-1100/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-1200/README.md b/checkpoint-1200/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-1200/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-1200/adapter_config.json b/checkpoint-1200/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-1200/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-1200/adapter_model.bin b/checkpoint-1200/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..f492dfb47359bd7e4f553ab1b632cf467f344dc7
--- /dev/null
+++ b/checkpoint-1200/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:12968dd3374591da9108d33e9902eb7d308edc5381a57c32c4be86471025fe20
+size 39409357
diff --git a/checkpoint-1200/optimizer.pt b/checkpoint-1200/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..a3a546e6f00bc7cb34b34a4f33e65d7319d76107
--- /dev/null
+++ b/checkpoint-1200/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3441b8e2f6f31ce90cfae67eb2109bfdede30ba5ad2d8e44cbbe7c61214bd3f2
+size 78844421
diff --git a/checkpoint-1200/rng_state.pth b/checkpoint-1200/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..d6f0f64e8115144f892f2b588cbb58afc5152fb1
--- /dev/null
+++ b/checkpoint-1200/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:91db4cb8b4c509d633633237f7fc5865ed0256714ee41edb2e45f150f9672a02
+size 14575
diff --git a/checkpoint-1200/scheduler.pt b/checkpoint-1200/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..ed07a9e856109c9d7e31ac59f4f05457bb0acf1d
--- /dev/null
+++ b/checkpoint-1200/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d75dac35039ac49dba103114c80c66a12f59c7fd524db7ff27a94928098ce8c4
+size 627
diff --git a/checkpoint-1200/special_tokens_map.json b/checkpoint-1200/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-1200/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-1200/tokenizer.json b/checkpoint-1200/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-1200/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-1200/tokenizer_config.json b/checkpoint-1200/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-1200/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-1200/trainer_state.json b/checkpoint-1200/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..d0b248e07cbc7873408d50262257f874a7bd448b
--- /dev/null
+++ b/checkpoint-1200/trainer_state.json
@@ -0,0 +1,7219 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 30.236220472440944,
+ "eval_steps": 500,
+ "global_step": 1200,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ },
+ {
+ "epoch": 10.1,
+ "learning_rate": 0.00033934294871794873,
+ "loss": 0.7635,
+ "step": 401
+ },
+ {
+ "epoch": 10.13,
+ "learning_rate": 0.00033894230769230773,
+ "loss": 0.7713,
+ "step": 402
+ },
+ {
+ "epoch": 10.15,
+ "learning_rate": 0.0003385416666666667,
+ "loss": 0.827,
+ "step": 403
+ },
+ {
+ "epoch": 10.18,
+ "learning_rate": 0.0003381410256410257,
+ "loss": 0.7844,
+ "step": 404
+ },
+ {
+ "epoch": 10.2,
+ "learning_rate": 0.00033774038461538463,
+ "loss": 0.7839,
+ "step": 405
+ },
+ {
+ "epoch": 10.23,
+ "learning_rate": 0.00033733974358974363,
+ "loss": 0.7702,
+ "step": 406
+ },
+ {
+ "epoch": 10.26,
+ "learning_rate": 0.0003369391025641026,
+ "loss": 0.7855,
+ "step": 407
+ },
+ {
+ "epoch": 10.28,
+ "learning_rate": 0.0003365384615384616,
+ "loss": 0.7669,
+ "step": 408
+ },
+ {
+ "epoch": 10.31,
+ "learning_rate": 0.0003361378205128205,
+ "loss": 0.8054,
+ "step": 409
+ },
+ {
+ "epoch": 10.33,
+ "learning_rate": 0.0003357371794871795,
+ "loss": 0.7792,
+ "step": 410
+ },
+ {
+ "epoch": 10.36,
+ "learning_rate": 0.0003353365384615384,
+ "loss": 0.7986,
+ "step": 411
+ },
+ {
+ "epoch": 10.38,
+ "learning_rate": 0.0003349358974358974,
+ "loss": 0.8269,
+ "step": 412
+ },
+ {
+ "epoch": 10.41,
+ "learning_rate": 0.00033453525641025637,
+ "loss": 0.7923,
+ "step": 413
+ },
+ {
+ "epoch": 10.43,
+ "learning_rate": 0.00033413461538461537,
+ "loss": 0.8242,
+ "step": 414
+ },
+ {
+ "epoch": 10.46,
+ "learning_rate": 0.0003337339743589743,
+ "loss": 0.8496,
+ "step": 415
+ },
+ {
+ "epoch": 10.48,
+ "learning_rate": 0.0003333333333333333,
+ "loss": 0.8236,
+ "step": 416
+ },
+ {
+ "epoch": 10.51,
+ "learning_rate": 0.0003329326923076923,
+ "loss": 0.8685,
+ "step": 417
+ },
+ {
+ "epoch": 10.53,
+ "learning_rate": 0.00033253205128205127,
+ "loss": 0.8542,
+ "step": 418
+ },
+ {
+ "epoch": 10.56,
+ "learning_rate": 0.0003321314102564103,
+ "loss": 0.8185,
+ "step": 419
+ },
+ {
+ "epoch": 10.58,
+ "learning_rate": 0.0003317307692307692,
+ "loss": 0.8477,
+ "step": 420
+ },
+ {
+ "epoch": 10.61,
+ "learning_rate": 0.0003313301282051282,
+ "loss": 0.8372,
+ "step": 421
+ },
+ {
+ "epoch": 10.63,
+ "learning_rate": 0.00033092948717948717,
+ "loss": 0.8731,
+ "step": 422
+ },
+ {
+ "epoch": 10.66,
+ "learning_rate": 0.00033052884615384617,
+ "loss": 0.8322,
+ "step": 423
+ },
+ {
+ "epoch": 10.68,
+ "learning_rate": 0.0003301282051282051,
+ "loss": 0.8514,
+ "step": 424
+ },
+ {
+ "epoch": 10.71,
+ "learning_rate": 0.0003297275641025641,
+ "loss": 0.8564,
+ "step": 425
+ },
+ {
+ "epoch": 10.73,
+ "learning_rate": 0.00032932692307692307,
+ "loss": 0.8278,
+ "step": 426
+ },
+ {
+ "epoch": 10.76,
+ "learning_rate": 0.00032892628205128207,
+ "loss": 0.8408,
+ "step": 427
+ },
+ {
+ "epoch": 10.78,
+ "learning_rate": 0.000328525641025641,
+ "loss": 0.8438,
+ "step": 428
+ },
+ {
+ "epoch": 10.81,
+ "learning_rate": 0.000328125,
+ "loss": 0.8862,
+ "step": 429
+ },
+ {
+ "epoch": 10.83,
+ "learning_rate": 0.00032772435897435897,
+ "loss": 0.8631,
+ "step": 430
+ },
+ {
+ "epoch": 10.86,
+ "learning_rate": 0.00032732371794871797,
+ "loss": 0.883,
+ "step": 431
+ },
+ {
+ "epoch": 10.89,
+ "learning_rate": 0.0003269230769230769,
+ "loss": 0.8644,
+ "step": 432
+ },
+ {
+ "epoch": 10.91,
+ "learning_rate": 0.0003265224358974359,
+ "loss": 0.8907,
+ "step": 433
+ },
+ {
+ "epoch": 10.94,
+ "learning_rate": 0.00032612179487179487,
+ "loss": 0.9141,
+ "step": 434
+ },
+ {
+ "epoch": 10.96,
+ "learning_rate": 0.00032572115384615387,
+ "loss": 0.8877,
+ "step": 435
+ },
+ {
+ "epoch": 10.99,
+ "learning_rate": 0.0003253205128205128,
+ "loss": 0.8521,
+ "step": 436
+ },
+ {
+ "epoch": 11.01,
+ "learning_rate": 0.0003249198717948718,
+ "loss": 0.7842,
+ "step": 437
+ },
+ {
+ "epoch": 11.04,
+ "learning_rate": 0.00032451923076923077,
+ "loss": 0.6516,
+ "step": 438
+ },
+ {
+ "epoch": 11.06,
+ "learning_rate": 0.00032411858974358977,
+ "loss": 0.6549,
+ "step": 439
+ },
+ {
+ "epoch": 11.09,
+ "learning_rate": 0.0003237179487179487,
+ "loss": 0.6736,
+ "step": 440
+ },
+ {
+ "epoch": 11.11,
+ "learning_rate": 0.0003233173076923077,
+ "loss": 0.6625,
+ "step": 441
+ },
+ {
+ "epoch": 11.14,
+ "learning_rate": 0.0003229166666666667,
+ "loss": 0.6466,
+ "step": 442
+ },
+ {
+ "epoch": 11.16,
+ "learning_rate": 0.00032251602564102567,
+ "loss": 0.6362,
+ "step": 443
+ },
+ {
+ "epoch": 11.19,
+ "learning_rate": 0.00032211538461538467,
+ "loss": 0.674,
+ "step": 444
+ },
+ {
+ "epoch": 11.21,
+ "learning_rate": 0.0003217147435897436,
+ "loss": 0.6509,
+ "step": 445
+ },
+ {
+ "epoch": 11.24,
+ "learning_rate": 0.0003213141025641026,
+ "loss": 0.618,
+ "step": 446
+ },
+ {
+ "epoch": 11.26,
+ "learning_rate": 0.00032091346153846156,
+ "loss": 0.6654,
+ "step": 447
+ },
+ {
+ "epoch": 11.29,
+ "learning_rate": 0.00032051282051282057,
+ "loss": 0.6632,
+ "step": 448
+ },
+ {
+ "epoch": 11.31,
+ "learning_rate": 0.00032011217948717946,
+ "loss": 0.6599,
+ "step": 449
+ },
+ {
+ "epoch": 11.34,
+ "learning_rate": 0.00031971153846153846,
+ "loss": 0.6537,
+ "step": 450
+ },
+ {
+ "epoch": 11.36,
+ "learning_rate": 0.0003193108974358974,
+ "loss": 0.678,
+ "step": 451
+ },
+ {
+ "epoch": 11.39,
+ "learning_rate": 0.0003189102564102564,
+ "loss": 0.684,
+ "step": 452
+ },
+ {
+ "epoch": 11.41,
+ "learning_rate": 0.00031850961538461536,
+ "loss": 0.6419,
+ "step": 453
+ },
+ {
+ "epoch": 11.44,
+ "learning_rate": 0.00031810897435897436,
+ "loss": 0.6614,
+ "step": 454
+ },
+ {
+ "epoch": 11.46,
+ "learning_rate": 0.0003177083333333333,
+ "loss": 0.7037,
+ "step": 455
+ },
+ {
+ "epoch": 11.49,
+ "learning_rate": 0.0003173076923076923,
+ "loss": 0.6944,
+ "step": 456
+ },
+ {
+ "epoch": 11.51,
+ "learning_rate": 0.00031690705128205126,
+ "loss": 0.6818,
+ "step": 457
+ },
+ {
+ "epoch": 11.54,
+ "learning_rate": 0.00031650641025641026,
+ "loss": 0.692,
+ "step": 458
+ },
+ {
+ "epoch": 11.57,
+ "learning_rate": 0.0003161057692307692,
+ "loss": 0.7028,
+ "step": 459
+ },
+ {
+ "epoch": 11.59,
+ "learning_rate": 0.0003157051282051282,
+ "loss": 0.7009,
+ "step": 460
+ },
+ {
+ "epoch": 11.62,
+ "learning_rate": 0.00031530448717948716,
+ "loss": 0.7325,
+ "step": 461
+ },
+ {
+ "epoch": 11.64,
+ "learning_rate": 0.00031490384615384616,
+ "loss": 0.6738,
+ "step": 462
+ },
+ {
+ "epoch": 11.67,
+ "learning_rate": 0.0003145032051282051,
+ "loss": 0.7008,
+ "step": 463
+ },
+ {
+ "epoch": 11.69,
+ "learning_rate": 0.0003141025641025641,
+ "loss": 0.6838,
+ "step": 464
+ },
+ {
+ "epoch": 11.72,
+ "learning_rate": 0.00031370192307692306,
+ "loss": 0.726,
+ "step": 465
+ },
+ {
+ "epoch": 11.74,
+ "learning_rate": 0.00031330128205128206,
+ "loss": 0.7167,
+ "step": 466
+ },
+ {
+ "epoch": 11.77,
+ "learning_rate": 0.000312900641025641,
+ "loss": 0.7236,
+ "step": 467
+ },
+ {
+ "epoch": 11.79,
+ "learning_rate": 0.0003125,
+ "loss": 0.7076,
+ "step": 468
+ },
+ {
+ "epoch": 11.82,
+ "learning_rate": 0.000312099358974359,
+ "loss": 0.7415,
+ "step": 469
+ },
+ {
+ "epoch": 11.84,
+ "learning_rate": 0.00031169871794871796,
+ "loss": 0.7313,
+ "step": 470
+ },
+ {
+ "epoch": 11.87,
+ "learning_rate": 0.00031129807692307696,
+ "loss": 0.7403,
+ "step": 471
+ },
+ {
+ "epoch": 11.89,
+ "learning_rate": 0.0003108974358974359,
+ "loss": 0.7145,
+ "step": 472
+ },
+ {
+ "epoch": 11.92,
+ "learning_rate": 0.0003104967948717949,
+ "loss": 0.7532,
+ "step": 473
+ },
+ {
+ "epoch": 11.94,
+ "learning_rate": 0.00031009615384615385,
+ "loss": 0.7326,
+ "step": 474
+ },
+ {
+ "epoch": 11.97,
+ "learning_rate": 0.00030969551282051286,
+ "loss": 0.7699,
+ "step": 475
+ },
+ {
+ "epoch": 11.99,
+ "learning_rate": 0.0003092948717948718,
+ "loss": 0.7251,
+ "step": 476
+ },
+ {
+ "epoch": 12.02,
+ "learning_rate": 0.0003088942307692308,
+ "loss": 0.5773,
+ "step": 477
+ },
+ {
+ "epoch": 12.04,
+ "learning_rate": 0.00030849358974358975,
+ "loss": 0.5288,
+ "step": 478
+ },
+ {
+ "epoch": 12.07,
+ "learning_rate": 0.00030809294871794876,
+ "loss": 0.5312,
+ "step": 479
+ },
+ {
+ "epoch": 12.09,
+ "learning_rate": 0.0003076923076923077,
+ "loss": 0.5445,
+ "step": 480
+ },
+ {
+ "epoch": 12.12,
+ "learning_rate": 0.0003072916666666667,
+ "loss": 0.5218,
+ "step": 481
+ },
+ {
+ "epoch": 12.14,
+ "learning_rate": 0.00030689102564102565,
+ "loss": 0.5311,
+ "step": 482
+ },
+ {
+ "epoch": 12.17,
+ "learning_rate": 0.00030649038461538465,
+ "loss": 0.5541,
+ "step": 483
+ },
+ {
+ "epoch": 12.2,
+ "learning_rate": 0.0003060897435897436,
+ "loss": 0.5242,
+ "step": 484
+ },
+ {
+ "epoch": 12.22,
+ "learning_rate": 0.0003056891025641026,
+ "loss": 0.5302,
+ "step": 485
+ },
+ {
+ "epoch": 12.25,
+ "learning_rate": 0.00030528846153846155,
+ "loss": 0.536,
+ "step": 486
+ },
+ {
+ "epoch": 12.27,
+ "learning_rate": 0.00030488782051282055,
+ "loss": 0.5545,
+ "step": 487
+ },
+ {
+ "epoch": 12.3,
+ "learning_rate": 0.00030448717948717945,
+ "loss": 0.5476,
+ "step": 488
+ },
+ {
+ "epoch": 12.32,
+ "learning_rate": 0.00030408653846153845,
+ "loss": 0.5528,
+ "step": 489
+ },
+ {
+ "epoch": 12.35,
+ "learning_rate": 0.0003036858974358974,
+ "loss": 0.5633,
+ "step": 490
+ },
+ {
+ "epoch": 12.37,
+ "learning_rate": 0.0003032852564102564,
+ "loss": 0.571,
+ "step": 491
+ },
+ {
+ "epoch": 12.4,
+ "learning_rate": 0.00030288461538461535,
+ "loss": 0.5637,
+ "step": 492
+ },
+ {
+ "epoch": 12.42,
+ "learning_rate": 0.00030248397435897435,
+ "loss": 0.5613,
+ "step": 493
+ },
+ {
+ "epoch": 12.45,
+ "learning_rate": 0.0003020833333333333,
+ "loss": 0.5999,
+ "step": 494
+ },
+ {
+ "epoch": 12.47,
+ "learning_rate": 0.0003016826923076923,
+ "loss": 0.5633,
+ "step": 495
+ },
+ {
+ "epoch": 12.5,
+ "learning_rate": 0.0003012820512820513,
+ "loss": 0.5811,
+ "step": 496
+ },
+ {
+ "epoch": 12.52,
+ "learning_rate": 0.00030088141025641025,
+ "loss": 0.5711,
+ "step": 497
+ },
+ {
+ "epoch": 12.55,
+ "learning_rate": 0.00030048076923076925,
+ "loss": 0.5669,
+ "step": 498
+ },
+ {
+ "epoch": 12.57,
+ "learning_rate": 0.0003000801282051282,
+ "loss": 0.5424,
+ "step": 499
+ },
+ {
+ "epoch": 12.6,
+ "learning_rate": 0.0002996794871794872,
+ "loss": 0.5592,
+ "step": 500
+ },
+ {
+ "epoch": 12.62,
+ "learning_rate": 0.00029927884615384614,
+ "loss": 0.5645,
+ "step": 501
+ },
+ {
+ "epoch": 12.65,
+ "learning_rate": 0.00029887820512820515,
+ "loss": 0.6089,
+ "step": 502
+ },
+ {
+ "epoch": 12.67,
+ "learning_rate": 0.0002984775641025641,
+ "loss": 0.57,
+ "step": 503
+ },
+ {
+ "epoch": 12.7,
+ "learning_rate": 0.0002980769230769231,
+ "loss": 0.57,
+ "step": 504
+ },
+ {
+ "epoch": 12.72,
+ "learning_rate": 0.00029767628205128204,
+ "loss": 0.5511,
+ "step": 505
+ },
+ {
+ "epoch": 12.75,
+ "learning_rate": 0.00029727564102564105,
+ "loss": 0.5765,
+ "step": 506
+ },
+ {
+ "epoch": 12.77,
+ "learning_rate": 0.000296875,
+ "loss": 0.5896,
+ "step": 507
+ },
+ {
+ "epoch": 12.8,
+ "learning_rate": 0.000296474358974359,
+ "loss": 0.6329,
+ "step": 508
+ },
+ {
+ "epoch": 12.83,
+ "learning_rate": 0.00029607371794871794,
+ "loss": 0.5682,
+ "step": 509
+ },
+ {
+ "epoch": 12.85,
+ "learning_rate": 0.00029567307692307694,
+ "loss": 0.58,
+ "step": 510
+ },
+ {
+ "epoch": 12.88,
+ "learning_rate": 0.0002952724358974359,
+ "loss": 0.6014,
+ "step": 511
+ },
+ {
+ "epoch": 12.9,
+ "learning_rate": 0.0002948717948717949,
+ "loss": 0.6082,
+ "step": 512
+ },
+ {
+ "epoch": 12.93,
+ "learning_rate": 0.00029447115384615384,
+ "loss": 0.6392,
+ "step": 513
+ },
+ {
+ "epoch": 12.95,
+ "learning_rate": 0.00029407051282051284,
+ "loss": 0.609,
+ "step": 514
+ },
+ {
+ "epoch": 12.98,
+ "learning_rate": 0.0002936698717948718,
+ "loss": 0.6247,
+ "step": 515
+ },
+ {
+ "epoch": 13.0,
+ "learning_rate": 0.0002932692307692308,
+ "loss": 0.5943,
+ "step": 516
+ },
+ {
+ "epoch": 13.03,
+ "learning_rate": 0.00029286858974358974,
+ "loss": 0.4499,
+ "step": 517
+ },
+ {
+ "epoch": 13.05,
+ "learning_rate": 0.00029246794871794874,
+ "loss": 0.4412,
+ "step": 518
+ },
+ {
+ "epoch": 13.08,
+ "learning_rate": 0.0002920673076923077,
+ "loss": 0.4409,
+ "step": 519
+ },
+ {
+ "epoch": 13.1,
+ "learning_rate": 0.0002916666666666667,
+ "loss": 0.4505,
+ "step": 520
+ },
+ {
+ "epoch": 13.13,
+ "learning_rate": 0.0002912660256410257,
+ "loss": 0.4489,
+ "step": 521
+ },
+ {
+ "epoch": 13.15,
+ "learning_rate": 0.00029086538461538464,
+ "loss": 0.4235,
+ "step": 522
+ },
+ {
+ "epoch": 13.18,
+ "learning_rate": 0.00029046474358974364,
+ "loss": 0.4276,
+ "step": 523
+ },
+ {
+ "epoch": 13.2,
+ "learning_rate": 0.0002900641025641026,
+ "loss": 0.4544,
+ "step": 524
+ },
+ {
+ "epoch": 13.23,
+ "learning_rate": 0.0002896634615384616,
+ "loss": 0.4625,
+ "step": 525
+ },
+ {
+ "epoch": 13.25,
+ "learning_rate": 0.00028926282051282054,
+ "loss": 0.4266,
+ "step": 526
+ },
+ {
+ "epoch": 13.28,
+ "learning_rate": 0.0002888621794871795,
+ "loss": 0.4459,
+ "step": 527
+ },
+ {
+ "epoch": 13.3,
+ "learning_rate": 0.00028846153846153843,
+ "loss": 0.4261,
+ "step": 528
+ },
+ {
+ "epoch": 13.33,
+ "learning_rate": 0.00028806089743589744,
+ "loss": 0.4376,
+ "step": 529
+ },
+ {
+ "epoch": 13.35,
+ "learning_rate": 0.0002876602564102564,
+ "loss": 0.4418,
+ "step": 530
+ },
+ {
+ "epoch": 13.38,
+ "learning_rate": 0.0002872596153846154,
+ "loss": 0.4667,
+ "step": 531
+ },
+ {
+ "epoch": 13.4,
+ "learning_rate": 0.00028685897435897433,
+ "loss": 0.4504,
+ "step": 532
+ },
+ {
+ "epoch": 13.43,
+ "learning_rate": 0.00028645833333333333,
+ "loss": 0.4691,
+ "step": 533
+ },
+ {
+ "epoch": 13.46,
+ "learning_rate": 0.0002860576923076923,
+ "loss": 0.4684,
+ "step": 534
+ },
+ {
+ "epoch": 13.48,
+ "learning_rate": 0.0002856570512820513,
+ "loss": 0.467,
+ "step": 535
+ },
+ {
+ "epoch": 13.51,
+ "learning_rate": 0.00028525641025641023,
+ "loss": 0.4716,
+ "step": 536
+ },
+ {
+ "epoch": 13.53,
+ "learning_rate": 0.00028485576923076923,
+ "loss": 0.4619,
+ "step": 537
+ },
+ {
+ "epoch": 13.56,
+ "learning_rate": 0.0002844551282051282,
+ "loss": 0.4775,
+ "step": 538
+ },
+ {
+ "epoch": 13.58,
+ "learning_rate": 0.0002840544871794872,
+ "loss": 0.4568,
+ "step": 539
+ },
+ {
+ "epoch": 13.61,
+ "learning_rate": 0.00028365384615384613,
+ "loss": 0.4763,
+ "step": 540
+ },
+ {
+ "epoch": 13.63,
+ "learning_rate": 0.00028325320512820513,
+ "loss": 0.466,
+ "step": 541
+ },
+ {
+ "epoch": 13.66,
+ "learning_rate": 0.0002828525641025641,
+ "loss": 0.4543,
+ "step": 542
+ },
+ {
+ "epoch": 13.68,
+ "learning_rate": 0.0002824519230769231,
+ "loss": 0.4666,
+ "step": 543
+ },
+ {
+ "epoch": 13.71,
+ "learning_rate": 0.00028205128205128203,
+ "loss": 0.489,
+ "step": 544
+ },
+ {
+ "epoch": 13.73,
+ "learning_rate": 0.00028165064102564103,
+ "loss": 0.491,
+ "step": 545
+ },
+ {
+ "epoch": 13.76,
+ "learning_rate": 0.00028125000000000003,
+ "loss": 0.4847,
+ "step": 546
+ },
+ {
+ "epoch": 13.78,
+ "learning_rate": 0.000280849358974359,
+ "loss": 0.4899,
+ "step": 547
+ },
+ {
+ "epoch": 13.81,
+ "learning_rate": 0.000280448717948718,
+ "loss": 0.5169,
+ "step": 548
+ },
+ {
+ "epoch": 13.83,
+ "learning_rate": 0.00028004807692307693,
+ "loss": 0.4968,
+ "step": 549
+ },
+ {
+ "epoch": 13.86,
+ "learning_rate": 0.00027964743589743593,
+ "loss": 0.4774,
+ "step": 550
+ },
+ {
+ "epoch": 13.88,
+ "learning_rate": 0.0002792467948717949,
+ "loss": 0.5105,
+ "step": 551
+ },
+ {
+ "epoch": 13.91,
+ "learning_rate": 0.0002788461538461539,
+ "loss": 0.4997,
+ "step": 552
+ },
+ {
+ "epoch": 13.93,
+ "learning_rate": 0.00027844551282051283,
+ "loss": 0.5088,
+ "step": 553
+ },
+ {
+ "epoch": 13.96,
+ "learning_rate": 0.00027804487179487183,
+ "loss": 0.5091,
+ "step": 554
+ },
+ {
+ "epoch": 13.98,
+ "learning_rate": 0.0002776442307692308,
+ "loss": 0.532,
+ "step": 555
+ },
+ {
+ "epoch": 14.01,
+ "learning_rate": 0.0002772435897435898,
+ "loss": 0.4487,
+ "step": 556
+ },
+ {
+ "epoch": 14.03,
+ "learning_rate": 0.00027684294871794873,
+ "loss": 0.3525,
+ "step": 557
+ },
+ {
+ "epoch": 14.06,
+ "learning_rate": 0.00027644230769230773,
+ "loss": 0.3715,
+ "step": 558
+ },
+ {
+ "epoch": 14.09,
+ "learning_rate": 0.0002760416666666667,
+ "loss": 0.3547,
+ "step": 559
+ },
+ {
+ "epoch": 14.11,
+ "learning_rate": 0.0002756410256410257,
+ "loss": 0.3862,
+ "step": 560
+ },
+ {
+ "epoch": 14.14,
+ "learning_rate": 0.0002752403846153846,
+ "loss": 0.3514,
+ "step": 561
+ },
+ {
+ "epoch": 14.16,
+ "learning_rate": 0.00027483974358974363,
+ "loss": 0.362,
+ "step": 562
+ },
+ {
+ "epoch": 14.19,
+ "learning_rate": 0.0002744391025641026,
+ "loss": 0.3457,
+ "step": 563
+ },
+ {
+ "epoch": 14.21,
+ "learning_rate": 0.0002740384615384616,
+ "loss": 0.3662,
+ "step": 564
+ },
+ {
+ "epoch": 14.24,
+ "learning_rate": 0.0002736378205128205,
+ "loss": 0.3746,
+ "step": 565
+ },
+ {
+ "epoch": 14.26,
+ "learning_rate": 0.00027323717948717947,
+ "loss": 0.3388,
+ "step": 566
+ },
+ {
+ "epoch": 14.29,
+ "learning_rate": 0.0002728365384615384,
+ "loss": 0.3824,
+ "step": 567
+ },
+ {
+ "epoch": 14.31,
+ "learning_rate": 0.0002724358974358974,
+ "loss": 0.3584,
+ "step": 568
+ },
+ {
+ "epoch": 14.34,
+ "learning_rate": 0.00027203525641025637,
+ "loss": 0.3544,
+ "step": 569
+ },
+ {
+ "epoch": 14.36,
+ "learning_rate": 0.00027163461538461537,
+ "loss": 0.3721,
+ "step": 570
+ },
+ {
+ "epoch": 14.39,
+ "learning_rate": 0.0002712339743589743,
+ "loss": 0.3769,
+ "step": 571
+ },
+ {
+ "epoch": 14.41,
+ "learning_rate": 0.0002708333333333333,
+ "loss": 0.3872,
+ "step": 572
+ },
+ {
+ "epoch": 14.44,
+ "learning_rate": 0.0002704326923076923,
+ "loss": 0.363,
+ "step": 573
+ },
+ {
+ "epoch": 14.46,
+ "learning_rate": 0.00027003205128205127,
+ "loss": 0.3842,
+ "step": 574
+ },
+ {
+ "epoch": 14.49,
+ "learning_rate": 0.00026963141025641027,
+ "loss": 0.3737,
+ "step": 575
+ },
+ {
+ "epoch": 14.51,
+ "learning_rate": 0.0002692307692307692,
+ "loss": 0.3981,
+ "step": 576
+ },
+ {
+ "epoch": 14.54,
+ "learning_rate": 0.0002688301282051282,
+ "loss": 0.3656,
+ "step": 577
+ },
+ {
+ "epoch": 14.56,
+ "learning_rate": 0.00026842948717948717,
+ "loss": 0.3852,
+ "step": 578
+ },
+ {
+ "epoch": 14.59,
+ "learning_rate": 0.00026802884615384617,
+ "loss": 0.3865,
+ "step": 579
+ },
+ {
+ "epoch": 14.61,
+ "learning_rate": 0.0002676282051282051,
+ "loss": 0.3747,
+ "step": 580
+ },
+ {
+ "epoch": 14.64,
+ "learning_rate": 0.0002672275641025641,
+ "loss": 0.3729,
+ "step": 581
+ },
+ {
+ "epoch": 14.66,
+ "learning_rate": 0.00026682692307692307,
+ "loss": 0.3785,
+ "step": 582
+ },
+ {
+ "epoch": 14.69,
+ "learning_rate": 0.00026642628205128207,
+ "loss": 0.4018,
+ "step": 583
+ },
+ {
+ "epoch": 14.71,
+ "learning_rate": 0.000266025641025641,
+ "loss": 0.3854,
+ "step": 584
+ },
+ {
+ "epoch": 14.74,
+ "learning_rate": 0.000265625,
+ "loss": 0.4161,
+ "step": 585
+ },
+ {
+ "epoch": 14.77,
+ "learning_rate": 0.00026522435897435897,
+ "loss": 0.4093,
+ "step": 586
+ },
+ {
+ "epoch": 14.79,
+ "learning_rate": 0.00026482371794871797,
+ "loss": 0.4126,
+ "step": 587
+ },
+ {
+ "epoch": 14.82,
+ "learning_rate": 0.0002644230769230769,
+ "loss": 0.3977,
+ "step": 588
+ },
+ {
+ "epoch": 14.84,
+ "learning_rate": 0.0002640224358974359,
+ "loss": 0.4286,
+ "step": 589
+ },
+ {
+ "epoch": 14.87,
+ "learning_rate": 0.00026362179487179487,
+ "loss": 0.4371,
+ "step": 590
+ },
+ {
+ "epoch": 14.89,
+ "learning_rate": 0.00026322115384615387,
+ "loss": 0.4066,
+ "step": 591
+ },
+ {
+ "epoch": 14.92,
+ "learning_rate": 0.0002628205128205128,
+ "loss": 0.433,
+ "step": 592
+ },
+ {
+ "epoch": 14.94,
+ "learning_rate": 0.0002624198717948718,
+ "loss": 0.4361,
+ "step": 593
+ },
+ {
+ "epoch": 14.97,
+ "learning_rate": 0.00026201923076923076,
+ "loss": 0.4276,
+ "step": 594
+ },
+ {
+ "epoch": 14.99,
+ "learning_rate": 0.00026161858974358977,
+ "loss": 0.4401,
+ "step": 595
+ },
+ {
+ "epoch": 15.02,
+ "learning_rate": 0.0002612179487179487,
+ "loss": 0.3514,
+ "step": 596
+ },
+ {
+ "epoch": 15.04,
+ "learning_rate": 0.0002608173076923077,
+ "loss": 0.306,
+ "step": 597
+ },
+ {
+ "epoch": 15.07,
+ "learning_rate": 0.0002604166666666667,
+ "loss": 0.2862,
+ "step": 598
+ },
+ {
+ "epoch": 15.09,
+ "learning_rate": 0.00026001602564102566,
+ "loss": 0.3042,
+ "step": 599
+ },
+ {
+ "epoch": 15.12,
+ "learning_rate": 0.00025961538461538467,
+ "loss": 0.2978,
+ "step": 600
+ },
+ {
+ "epoch": 15.14,
+ "learning_rate": 0.0002592147435897436,
+ "loss": 0.2975,
+ "step": 601
+ },
+ {
+ "epoch": 15.17,
+ "learning_rate": 0.0002588141025641026,
+ "loss": 0.3172,
+ "step": 602
+ },
+ {
+ "epoch": 15.19,
+ "learning_rate": 0.00025841346153846156,
+ "loss": 0.3046,
+ "step": 603
+ },
+ {
+ "epoch": 15.22,
+ "learning_rate": 0.00025801282051282057,
+ "loss": 0.3008,
+ "step": 604
+ },
+ {
+ "epoch": 15.24,
+ "learning_rate": 0.00025761217948717946,
+ "loss": 0.2956,
+ "step": 605
+ },
+ {
+ "epoch": 15.27,
+ "learning_rate": 0.00025721153846153846,
+ "loss": 0.3137,
+ "step": 606
+ },
+ {
+ "epoch": 15.29,
+ "learning_rate": 0.0002568108974358974,
+ "loss": 0.2961,
+ "step": 607
+ },
+ {
+ "epoch": 15.32,
+ "learning_rate": 0.0002564102564102564,
+ "loss": 0.3059,
+ "step": 608
+ },
+ {
+ "epoch": 15.34,
+ "learning_rate": 0.00025600961538461536,
+ "loss": 0.3077,
+ "step": 609
+ },
+ {
+ "epoch": 15.37,
+ "learning_rate": 0.00025560897435897436,
+ "loss": 0.3189,
+ "step": 610
+ },
+ {
+ "epoch": 15.4,
+ "learning_rate": 0.0002552083333333333,
+ "loss": 0.309,
+ "step": 611
+ },
+ {
+ "epoch": 15.42,
+ "learning_rate": 0.0002548076923076923,
+ "loss": 0.3324,
+ "step": 612
+ },
+ {
+ "epoch": 15.45,
+ "learning_rate": 0.00025440705128205126,
+ "loss": 0.3159,
+ "step": 613
+ },
+ {
+ "epoch": 15.47,
+ "learning_rate": 0.00025400641025641026,
+ "loss": 0.3188,
+ "step": 614
+ },
+ {
+ "epoch": 15.5,
+ "learning_rate": 0.0002536057692307692,
+ "loss": 0.3174,
+ "step": 615
+ },
+ {
+ "epoch": 15.52,
+ "learning_rate": 0.0002532051282051282,
+ "loss": 0.3074,
+ "step": 616
+ },
+ {
+ "epoch": 15.55,
+ "learning_rate": 0.00025280448717948716,
+ "loss": 0.3145,
+ "step": 617
+ },
+ {
+ "epoch": 15.57,
+ "learning_rate": 0.00025240384615384616,
+ "loss": 0.317,
+ "step": 618
+ },
+ {
+ "epoch": 15.6,
+ "learning_rate": 0.0002520032051282051,
+ "loss": 0.318,
+ "step": 619
+ },
+ {
+ "epoch": 15.62,
+ "learning_rate": 0.0002516025641025641,
+ "loss": 0.3236,
+ "step": 620
+ },
+ {
+ "epoch": 15.65,
+ "learning_rate": 0.00025120192307692305,
+ "loss": 0.3282,
+ "step": 621
+ },
+ {
+ "epoch": 15.67,
+ "learning_rate": 0.00025080128205128206,
+ "loss": 0.3394,
+ "step": 622
+ },
+ {
+ "epoch": 15.7,
+ "learning_rate": 0.000250400641025641,
+ "loss": 0.3278,
+ "step": 623
+ },
+ {
+ "epoch": 15.72,
+ "learning_rate": 0.00025,
+ "loss": 0.3255,
+ "step": 624
+ },
+ {
+ "epoch": 15.75,
+ "learning_rate": 0.000249599358974359,
+ "loss": 0.3256,
+ "step": 625
+ },
+ {
+ "epoch": 15.77,
+ "learning_rate": 0.00024919871794871795,
+ "loss": 0.3379,
+ "step": 626
+ },
+ {
+ "epoch": 15.8,
+ "learning_rate": 0.00024879807692307696,
+ "loss": 0.3384,
+ "step": 627
+ },
+ {
+ "epoch": 15.82,
+ "learning_rate": 0.0002483974358974359,
+ "loss": 0.3483,
+ "step": 628
+ },
+ {
+ "epoch": 15.85,
+ "learning_rate": 0.0002479967948717949,
+ "loss": 0.3478,
+ "step": 629
+ },
+ {
+ "epoch": 15.87,
+ "learning_rate": 0.00024759615384615385,
+ "loss": 0.337,
+ "step": 630
+ },
+ {
+ "epoch": 15.9,
+ "learning_rate": 0.00024719551282051286,
+ "loss": 0.3287,
+ "step": 631
+ },
+ {
+ "epoch": 15.92,
+ "learning_rate": 0.0002467948717948718,
+ "loss": 0.3478,
+ "step": 632
+ },
+ {
+ "epoch": 15.95,
+ "learning_rate": 0.0002463942307692308,
+ "loss": 0.3567,
+ "step": 633
+ },
+ {
+ "epoch": 15.97,
+ "learning_rate": 0.00024599358974358975,
+ "loss": 0.3359,
+ "step": 634
+ },
+ {
+ "epoch": 16.0,
+ "learning_rate": 0.0002455929487179487,
+ "loss": 0.3492,
+ "step": 635
+ },
+ {
+ "epoch": 16.03,
+ "learning_rate": 0.0002451923076923077,
+ "loss": 0.2497,
+ "step": 636
+ },
+ {
+ "epoch": 16.05,
+ "learning_rate": 0.00024479166666666665,
+ "loss": 0.2416,
+ "step": 637
+ },
+ {
+ "epoch": 16.08,
+ "learning_rate": 0.00024439102564102565,
+ "loss": 0.2576,
+ "step": 638
+ },
+ {
+ "epoch": 16.1,
+ "learning_rate": 0.00024399038461538463,
+ "loss": 0.243,
+ "step": 639
+ },
+ {
+ "epoch": 16.13,
+ "learning_rate": 0.0002435897435897436,
+ "loss": 0.2527,
+ "step": 640
+ },
+ {
+ "epoch": 16.15,
+ "learning_rate": 0.00024318910256410258,
+ "loss": 0.2606,
+ "step": 641
+ },
+ {
+ "epoch": 16.18,
+ "learning_rate": 0.00024278846153846155,
+ "loss": 0.2555,
+ "step": 642
+ },
+ {
+ "epoch": 16.2,
+ "learning_rate": 0.00024238782051282052,
+ "loss": 0.2581,
+ "step": 643
+ },
+ {
+ "epoch": 16.23,
+ "learning_rate": 0.0002419871794871795,
+ "loss": 0.2508,
+ "step": 644
+ },
+ {
+ "epoch": 16.25,
+ "learning_rate": 0.00024158653846153847,
+ "loss": 0.281,
+ "step": 645
+ },
+ {
+ "epoch": 16.28,
+ "learning_rate": 0.00024118589743589745,
+ "loss": 0.2626,
+ "step": 646
+ },
+ {
+ "epoch": 16.3,
+ "learning_rate": 0.00024078525641025642,
+ "loss": 0.2668,
+ "step": 647
+ },
+ {
+ "epoch": 16.33,
+ "learning_rate": 0.0002403846153846154,
+ "loss": 0.2675,
+ "step": 648
+ },
+ {
+ "epoch": 16.35,
+ "learning_rate": 0.00023998397435897437,
+ "loss": 0.2491,
+ "step": 649
+ },
+ {
+ "epoch": 16.38,
+ "learning_rate": 0.00023958333333333335,
+ "loss": 0.2459,
+ "step": 650
+ },
+ {
+ "epoch": 16.4,
+ "learning_rate": 0.00023918269230769232,
+ "loss": 0.2515,
+ "step": 651
+ },
+ {
+ "epoch": 16.43,
+ "learning_rate": 0.0002387820512820513,
+ "loss": 0.2612,
+ "step": 652
+ },
+ {
+ "epoch": 16.45,
+ "learning_rate": 0.00023838141025641027,
+ "loss": 0.2612,
+ "step": 653
+ },
+ {
+ "epoch": 16.48,
+ "learning_rate": 0.00023798076923076922,
+ "loss": 0.265,
+ "step": 654
+ },
+ {
+ "epoch": 16.5,
+ "learning_rate": 0.0002375801282051282,
+ "loss": 0.2634,
+ "step": 655
+ },
+ {
+ "epoch": 16.53,
+ "learning_rate": 0.00023717948717948717,
+ "loss": 0.2839,
+ "step": 656
+ },
+ {
+ "epoch": 16.55,
+ "learning_rate": 0.00023677884615384614,
+ "loss": 0.2595,
+ "step": 657
+ },
+ {
+ "epoch": 16.58,
+ "learning_rate": 0.00023637820512820512,
+ "loss": 0.2764,
+ "step": 658
+ },
+ {
+ "epoch": 16.6,
+ "learning_rate": 0.0002359775641025641,
+ "loss": 0.2553,
+ "step": 659
+ },
+ {
+ "epoch": 16.63,
+ "learning_rate": 0.00023557692307692307,
+ "loss": 0.2716,
+ "step": 660
+ },
+ {
+ "epoch": 16.66,
+ "learning_rate": 0.00023517628205128204,
+ "loss": 0.289,
+ "step": 661
+ },
+ {
+ "epoch": 16.68,
+ "learning_rate": 0.00023477564102564102,
+ "loss": 0.2483,
+ "step": 662
+ },
+ {
+ "epoch": 16.71,
+ "learning_rate": 0.000234375,
+ "loss": 0.269,
+ "step": 663
+ },
+ {
+ "epoch": 16.73,
+ "learning_rate": 0.000233974358974359,
+ "loss": 0.2738,
+ "step": 664
+ },
+ {
+ "epoch": 16.76,
+ "learning_rate": 0.00023357371794871797,
+ "loss": 0.2646,
+ "step": 665
+ },
+ {
+ "epoch": 16.78,
+ "learning_rate": 0.00023317307692307694,
+ "loss": 0.2793,
+ "step": 666
+ },
+ {
+ "epoch": 16.81,
+ "learning_rate": 0.00023277243589743592,
+ "loss": 0.2666,
+ "step": 667
+ },
+ {
+ "epoch": 16.83,
+ "learning_rate": 0.0002323717948717949,
+ "loss": 0.2799,
+ "step": 668
+ },
+ {
+ "epoch": 16.86,
+ "learning_rate": 0.00023197115384615387,
+ "loss": 0.2677,
+ "step": 669
+ },
+ {
+ "epoch": 16.88,
+ "learning_rate": 0.00023157051282051284,
+ "loss": 0.2889,
+ "step": 670
+ },
+ {
+ "epoch": 16.91,
+ "learning_rate": 0.00023116987179487182,
+ "loss": 0.2869,
+ "step": 671
+ },
+ {
+ "epoch": 16.93,
+ "learning_rate": 0.0002307692307692308,
+ "loss": 0.2814,
+ "step": 672
+ },
+ {
+ "epoch": 16.96,
+ "learning_rate": 0.00023036858974358974,
+ "loss": 0.2881,
+ "step": 673
+ },
+ {
+ "epoch": 16.98,
+ "learning_rate": 0.0002299679487179487,
+ "loss": 0.2777,
+ "step": 674
+ },
+ {
+ "epoch": 17.01,
+ "learning_rate": 0.0002295673076923077,
+ "loss": 0.2713,
+ "step": 675
+ },
+ {
+ "epoch": 17.03,
+ "learning_rate": 0.00022916666666666666,
+ "loss": 0.2176,
+ "step": 676
+ },
+ {
+ "epoch": 17.06,
+ "learning_rate": 0.00022876602564102564,
+ "loss": 0.2032,
+ "step": 677
+ },
+ {
+ "epoch": 17.08,
+ "learning_rate": 0.0002283653846153846,
+ "loss": 0.2207,
+ "step": 678
+ },
+ {
+ "epoch": 17.11,
+ "learning_rate": 0.0002279647435897436,
+ "loss": 0.2183,
+ "step": 679
+ },
+ {
+ "epoch": 17.13,
+ "learning_rate": 0.00022756410256410256,
+ "loss": 0.2238,
+ "step": 680
+ },
+ {
+ "epoch": 17.16,
+ "learning_rate": 0.00022716346153846154,
+ "loss": 0.2126,
+ "step": 681
+ },
+ {
+ "epoch": 17.18,
+ "learning_rate": 0.0002267628205128205,
+ "loss": 0.2104,
+ "step": 682
+ },
+ {
+ "epoch": 17.21,
+ "learning_rate": 0.00022636217948717949,
+ "loss": 0.2151,
+ "step": 683
+ },
+ {
+ "epoch": 17.23,
+ "learning_rate": 0.00022596153846153846,
+ "loss": 0.2093,
+ "step": 684
+ },
+ {
+ "epoch": 17.26,
+ "learning_rate": 0.00022556089743589743,
+ "loss": 0.2308,
+ "step": 685
+ },
+ {
+ "epoch": 17.29,
+ "learning_rate": 0.0002251602564102564,
+ "loss": 0.2276,
+ "step": 686
+ },
+ {
+ "epoch": 17.31,
+ "learning_rate": 0.00022475961538461538,
+ "loss": 0.2328,
+ "step": 687
+ },
+ {
+ "epoch": 17.34,
+ "learning_rate": 0.00022435897435897436,
+ "loss": 0.2076,
+ "step": 688
+ },
+ {
+ "epoch": 17.36,
+ "learning_rate": 0.00022395833333333336,
+ "loss": 0.2125,
+ "step": 689
+ },
+ {
+ "epoch": 17.39,
+ "learning_rate": 0.00022355769230769234,
+ "loss": 0.2181,
+ "step": 690
+ },
+ {
+ "epoch": 17.41,
+ "learning_rate": 0.0002231570512820513,
+ "loss": 0.2153,
+ "step": 691
+ },
+ {
+ "epoch": 17.44,
+ "learning_rate": 0.00022275641025641028,
+ "loss": 0.2122,
+ "step": 692
+ },
+ {
+ "epoch": 17.46,
+ "learning_rate": 0.00022235576923076923,
+ "loss": 0.2262,
+ "step": 693
+ },
+ {
+ "epoch": 17.49,
+ "learning_rate": 0.0002219551282051282,
+ "loss": 0.2206,
+ "step": 694
+ },
+ {
+ "epoch": 17.51,
+ "learning_rate": 0.00022155448717948718,
+ "loss": 0.2223,
+ "step": 695
+ },
+ {
+ "epoch": 17.54,
+ "learning_rate": 0.00022115384615384616,
+ "loss": 0.2306,
+ "step": 696
+ },
+ {
+ "epoch": 17.56,
+ "learning_rate": 0.00022075320512820513,
+ "loss": 0.213,
+ "step": 697
+ },
+ {
+ "epoch": 17.59,
+ "learning_rate": 0.0002203525641025641,
+ "loss": 0.2198,
+ "step": 698
+ },
+ {
+ "epoch": 17.61,
+ "learning_rate": 0.00021995192307692308,
+ "loss": 0.2299,
+ "step": 699
+ },
+ {
+ "epoch": 17.64,
+ "learning_rate": 0.00021955128205128206,
+ "loss": 0.2213,
+ "step": 700
+ },
+ {
+ "epoch": 17.66,
+ "learning_rate": 0.00021915064102564103,
+ "loss": 0.2183,
+ "step": 701
+ },
+ {
+ "epoch": 17.69,
+ "learning_rate": 0.00021875,
+ "loss": 0.2301,
+ "step": 702
+ },
+ {
+ "epoch": 17.71,
+ "learning_rate": 0.00021834935897435898,
+ "loss": 0.2348,
+ "step": 703
+ },
+ {
+ "epoch": 17.74,
+ "learning_rate": 0.00021794871794871795,
+ "loss": 0.229,
+ "step": 704
+ },
+ {
+ "epoch": 17.76,
+ "learning_rate": 0.00021754807692307693,
+ "loss": 0.2397,
+ "step": 705
+ },
+ {
+ "epoch": 17.79,
+ "learning_rate": 0.0002171474358974359,
+ "loss": 0.2254,
+ "step": 706
+ },
+ {
+ "epoch": 17.81,
+ "learning_rate": 0.00021674679487179488,
+ "loss": 0.2306,
+ "step": 707
+ },
+ {
+ "epoch": 17.84,
+ "learning_rate": 0.00021634615384615385,
+ "loss": 0.2388,
+ "step": 708
+ },
+ {
+ "epoch": 17.86,
+ "learning_rate": 0.00021594551282051283,
+ "loss": 0.2374,
+ "step": 709
+ },
+ {
+ "epoch": 17.89,
+ "learning_rate": 0.0002155448717948718,
+ "loss": 0.2412,
+ "step": 710
+ },
+ {
+ "epoch": 17.91,
+ "learning_rate": 0.00021514423076923078,
+ "loss": 0.2203,
+ "step": 711
+ },
+ {
+ "epoch": 17.94,
+ "learning_rate": 0.00021474358974358972,
+ "loss": 0.2359,
+ "step": 712
+ },
+ {
+ "epoch": 17.97,
+ "learning_rate": 0.0002143429487179487,
+ "loss": 0.2343,
+ "step": 713
+ },
+ {
+ "epoch": 17.99,
+ "learning_rate": 0.00021394230769230767,
+ "loss": 0.2397,
+ "step": 714
+ },
+ {
+ "epoch": 18.02,
+ "learning_rate": 0.00021354166666666665,
+ "loss": 0.2055,
+ "step": 715
+ },
+ {
+ "epoch": 18.04,
+ "learning_rate": 0.00021314102564102565,
+ "loss": 0.1723,
+ "step": 716
+ },
+ {
+ "epoch": 18.07,
+ "learning_rate": 0.00021274038461538463,
+ "loss": 0.1782,
+ "step": 717
+ },
+ {
+ "epoch": 18.09,
+ "learning_rate": 0.0002123397435897436,
+ "loss": 0.1813,
+ "step": 718
+ },
+ {
+ "epoch": 18.12,
+ "learning_rate": 0.00021193910256410257,
+ "loss": 0.1834,
+ "step": 719
+ },
+ {
+ "epoch": 18.14,
+ "learning_rate": 0.00021153846153846155,
+ "loss": 0.1759,
+ "step": 720
+ },
+ {
+ "epoch": 18.17,
+ "learning_rate": 0.00021113782051282052,
+ "loss": 0.1855,
+ "step": 721
+ },
+ {
+ "epoch": 18.19,
+ "learning_rate": 0.0002107371794871795,
+ "loss": 0.1756,
+ "step": 722
+ },
+ {
+ "epoch": 18.22,
+ "learning_rate": 0.00021033653846153847,
+ "loss": 0.1783,
+ "step": 723
+ },
+ {
+ "epoch": 18.24,
+ "learning_rate": 0.00020993589743589745,
+ "loss": 0.1899,
+ "step": 724
+ },
+ {
+ "epoch": 18.27,
+ "learning_rate": 0.00020953525641025642,
+ "loss": 0.1817,
+ "step": 725
+ },
+ {
+ "epoch": 18.29,
+ "learning_rate": 0.0002091346153846154,
+ "loss": 0.1856,
+ "step": 726
+ },
+ {
+ "epoch": 18.32,
+ "learning_rate": 0.00020873397435897437,
+ "loss": 0.1849,
+ "step": 727
+ },
+ {
+ "epoch": 18.34,
+ "learning_rate": 0.00020833333333333335,
+ "loss": 0.1758,
+ "step": 728
+ },
+ {
+ "epoch": 18.37,
+ "learning_rate": 0.00020793269230769232,
+ "loss": 0.1907,
+ "step": 729
+ },
+ {
+ "epoch": 18.39,
+ "learning_rate": 0.0002075320512820513,
+ "loss": 0.1836,
+ "step": 730
+ },
+ {
+ "epoch": 18.42,
+ "learning_rate": 0.00020713141025641027,
+ "loss": 0.1828,
+ "step": 731
+ },
+ {
+ "epoch": 18.44,
+ "learning_rate": 0.00020673076923076922,
+ "loss": 0.1916,
+ "step": 732
+ },
+ {
+ "epoch": 18.47,
+ "learning_rate": 0.0002063301282051282,
+ "loss": 0.1849,
+ "step": 733
+ },
+ {
+ "epoch": 18.49,
+ "learning_rate": 0.00020592948717948717,
+ "loss": 0.1929,
+ "step": 734
+ },
+ {
+ "epoch": 18.52,
+ "learning_rate": 0.00020552884615384614,
+ "loss": 0.1779,
+ "step": 735
+ },
+ {
+ "epoch": 18.54,
+ "learning_rate": 0.00020512820512820512,
+ "loss": 0.1805,
+ "step": 736
+ },
+ {
+ "epoch": 18.57,
+ "learning_rate": 0.0002047275641025641,
+ "loss": 0.1891,
+ "step": 737
+ },
+ {
+ "epoch": 18.6,
+ "learning_rate": 0.00020432692307692307,
+ "loss": 0.1932,
+ "step": 738
+ },
+ {
+ "epoch": 18.62,
+ "learning_rate": 0.00020392628205128204,
+ "loss": 0.1923,
+ "step": 739
+ },
+ {
+ "epoch": 18.65,
+ "learning_rate": 0.00020352564102564102,
+ "loss": 0.1882,
+ "step": 740
+ },
+ {
+ "epoch": 18.67,
+ "learning_rate": 0.00020312500000000002,
+ "loss": 0.2012,
+ "step": 741
+ },
+ {
+ "epoch": 18.7,
+ "learning_rate": 0.000202724358974359,
+ "loss": 0.1963,
+ "step": 742
+ },
+ {
+ "epoch": 18.72,
+ "learning_rate": 0.00020232371794871797,
+ "loss": 0.202,
+ "step": 743
+ },
+ {
+ "epoch": 18.75,
+ "learning_rate": 0.00020192307692307694,
+ "loss": 0.1996,
+ "step": 744
+ },
+ {
+ "epoch": 18.77,
+ "learning_rate": 0.00020152243589743592,
+ "loss": 0.2056,
+ "step": 745
+ },
+ {
+ "epoch": 18.8,
+ "learning_rate": 0.0002011217948717949,
+ "loss": 0.1946,
+ "step": 746
+ },
+ {
+ "epoch": 18.82,
+ "learning_rate": 0.00020072115384615387,
+ "loss": 0.1954,
+ "step": 747
+ },
+ {
+ "epoch": 18.85,
+ "learning_rate": 0.00020032051282051284,
+ "loss": 0.1966,
+ "step": 748
+ },
+ {
+ "epoch": 18.87,
+ "learning_rate": 0.00019991987179487182,
+ "loss": 0.2077,
+ "step": 749
+ },
+ {
+ "epoch": 18.9,
+ "learning_rate": 0.0001995192307692308,
+ "loss": 0.2,
+ "step": 750
+ },
+ {
+ "epoch": 18.92,
+ "learning_rate": 0.00019911858974358974,
+ "loss": 0.2092,
+ "step": 751
+ },
+ {
+ "epoch": 18.95,
+ "learning_rate": 0.0001987179487179487,
+ "loss": 0.2023,
+ "step": 752
+ },
+ {
+ "epoch": 18.97,
+ "learning_rate": 0.0001983173076923077,
+ "loss": 0.2056,
+ "step": 753
+ },
+ {
+ "epoch": 19.0,
+ "learning_rate": 0.00019791666666666666,
+ "loss": 0.2059,
+ "step": 754
+ },
+ {
+ "epoch": 19.02,
+ "learning_rate": 0.00019751602564102564,
+ "loss": 0.1581,
+ "step": 755
+ },
+ {
+ "epoch": 19.05,
+ "learning_rate": 0.0001971153846153846,
+ "loss": 0.155,
+ "step": 756
+ },
+ {
+ "epoch": 19.07,
+ "learning_rate": 0.00019671474358974359,
+ "loss": 0.156,
+ "step": 757
+ },
+ {
+ "epoch": 19.1,
+ "learning_rate": 0.00019631410256410256,
+ "loss": 0.161,
+ "step": 758
+ },
+ {
+ "epoch": 19.12,
+ "learning_rate": 0.00019591346153846154,
+ "loss": 0.1608,
+ "step": 759
+ },
+ {
+ "epoch": 19.15,
+ "learning_rate": 0.0001955128205128205,
+ "loss": 0.1576,
+ "step": 760
+ },
+ {
+ "epoch": 19.17,
+ "learning_rate": 0.00019511217948717948,
+ "loss": 0.1539,
+ "step": 761
+ },
+ {
+ "epoch": 19.2,
+ "learning_rate": 0.00019471153846153846,
+ "loss": 0.1549,
+ "step": 762
+ },
+ {
+ "epoch": 19.23,
+ "learning_rate": 0.00019431089743589743,
+ "loss": 0.1593,
+ "step": 763
+ },
+ {
+ "epoch": 19.25,
+ "learning_rate": 0.0001939102564102564,
+ "loss": 0.162,
+ "step": 764
+ },
+ {
+ "epoch": 19.28,
+ "learning_rate": 0.00019350961538461538,
+ "loss": 0.1591,
+ "step": 765
+ },
+ {
+ "epoch": 19.3,
+ "learning_rate": 0.00019310897435897436,
+ "loss": 0.1569,
+ "step": 766
+ },
+ {
+ "epoch": 19.33,
+ "learning_rate": 0.00019270833333333336,
+ "loss": 0.1628,
+ "step": 767
+ },
+ {
+ "epoch": 19.35,
+ "learning_rate": 0.00019230769230769233,
+ "loss": 0.1529,
+ "step": 768
+ },
+ {
+ "epoch": 19.38,
+ "learning_rate": 0.0001919070512820513,
+ "loss": 0.1606,
+ "step": 769
+ },
+ {
+ "epoch": 19.4,
+ "learning_rate": 0.00019150641025641028,
+ "loss": 0.1564,
+ "step": 770
+ },
+ {
+ "epoch": 19.43,
+ "learning_rate": 0.00019110576923076923,
+ "loss": 0.1599,
+ "step": 771
+ },
+ {
+ "epoch": 19.45,
+ "learning_rate": 0.0001907051282051282,
+ "loss": 0.1653,
+ "step": 772
+ },
+ {
+ "epoch": 19.48,
+ "learning_rate": 0.00019030448717948718,
+ "loss": 0.1587,
+ "step": 773
+ },
+ {
+ "epoch": 19.5,
+ "learning_rate": 0.00018990384615384616,
+ "loss": 0.1683,
+ "step": 774
+ },
+ {
+ "epoch": 19.53,
+ "learning_rate": 0.00018950320512820513,
+ "loss": 0.157,
+ "step": 775
+ },
+ {
+ "epoch": 19.55,
+ "learning_rate": 0.0001891025641025641,
+ "loss": 0.1588,
+ "step": 776
+ },
+ {
+ "epoch": 19.58,
+ "learning_rate": 0.00018870192307692308,
+ "loss": 0.1709,
+ "step": 777
+ },
+ {
+ "epoch": 19.6,
+ "learning_rate": 0.00018830128205128205,
+ "loss": 0.1635,
+ "step": 778
+ },
+ {
+ "epoch": 19.63,
+ "learning_rate": 0.00018790064102564103,
+ "loss": 0.1581,
+ "step": 779
+ },
+ {
+ "epoch": 19.65,
+ "learning_rate": 0.0001875,
+ "loss": 0.1664,
+ "step": 780
+ },
+ {
+ "epoch": 19.68,
+ "learning_rate": 0.00018709935897435898,
+ "loss": 0.1712,
+ "step": 781
+ },
+ {
+ "epoch": 19.7,
+ "learning_rate": 0.00018669871794871795,
+ "loss": 0.165,
+ "step": 782
+ },
+ {
+ "epoch": 19.73,
+ "learning_rate": 0.00018629807692307693,
+ "loss": 0.1692,
+ "step": 783
+ },
+ {
+ "epoch": 19.75,
+ "learning_rate": 0.0001858974358974359,
+ "loss": 0.1658,
+ "step": 784
+ },
+ {
+ "epoch": 19.78,
+ "learning_rate": 0.00018549679487179488,
+ "loss": 0.1702,
+ "step": 785
+ },
+ {
+ "epoch": 19.8,
+ "learning_rate": 0.00018509615384615385,
+ "loss": 0.1672,
+ "step": 786
+ },
+ {
+ "epoch": 19.83,
+ "learning_rate": 0.00018469551282051283,
+ "loss": 0.1708,
+ "step": 787
+ },
+ {
+ "epoch": 19.86,
+ "learning_rate": 0.0001842948717948718,
+ "loss": 0.1705,
+ "step": 788
+ },
+ {
+ "epoch": 19.88,
+ "learning_rate": 0.00018389423076923078,
+ "loss": 0.1853,
+ "step": 789
+ },
+ {
+ "epoch": 19.91,
+ "learning_rate": 0.00018349358974358972,
+ "loss": 0.1701,
+ "step": 790
+ },
+ {
+ "epoch": 19.93,
+ "learning_rate": 0.0001830929487179487,
+ "loss": 0.1726,
+ "step": 791
+ },
+ {
+ "epoch": 19.96,
+ "learning_rate": 0.00018269230769230767,
+ "loss": 0.1829,
+ "step": 792
+ },
+ {
+ "epoch": 19.98,
+ "learning_rate": 0.00018229166666666665,
+ "loss": 0.1749,
+ "step": 793
+ },
+ {
+ "epoch": 20.01,
+ "learning_rate": 0.00018189102564102565,
+ "loss": 0.1559,
+ "step": 794
+ },
+ {
+ "epoch": 20.03,
+ "learning_rate": 0.00018149038461538462,
+ "loss": 0.1424,
+ "step": 795
+ },
+ {
+ "epoch": 20.06,
+ "learning_rate": 0.0001810897435897436,
+ "loss": 0.1386,
+ "step": 796
+ },
+ {
+ "epoch": 20.08,
+ "learning_rate": 0.00018068910256410257,
+ "loss": 0.1388,
+ "step": 797
+ },
+ {
+ "epoch": 20.11,
+ "learning_rate": 0.00018028846153846155,
+ "loss": 0.1385,
+ "step": 798
+ },
+ {
+ "epoch": 20.13,
+ "learning_rate": 0.00017988782051282052,
+ "loss": 0.1349,
+ "step": 799
+ },
+ {
+ "epoch": 20.16,
+ "learning_rate": 0.0001794871794871795,
+ "loss": 0.1357,
+ "step": 800
+ },
+ {
+ "epoch": 20.18,
+ "learning_rate": 0.00017908653846153847,
+ "loss": 0.1384,
+ "step": 801
+ },
+ {
+ "epoch": 20.21,
+ "learning_rate": 0.00017868589743589745,
+ "loss": 0.1425,
+ "step": 802
+ },
+ {
+ "epoch": 20.23,
+ "learning_rate": 0.00017828525641025642,
+ "loss": 0.1337,
+ "step": 803
+ },
+ {
+ "epoch": 20.26,
+ "learning_rate": 0.0001778846153846154,
+ "loss": 0.136,
+ "step": 804
+ },
+ {
+ "epoch": 20.28,
+ "learning_rate": 0.00017748397435897437,
+ "loss": 0.1382,
+ "step": 805
+ },
+ {
+ "epoch": 20.31,
+ "learning_rate": 0.00017708333333333335,
+ "loss": 0.1396,
+ "step": 806
+ },
+ {
+ "epoch": 20.33,
+ "learning_rate": 0.00017668269230769232,
+ "loss": 0.1411,
+ "step": 807
+ },
+ {
+ "epoch": 20.36,
+ "learning_rate": 0.0001762820512820513,
+ "loss": 0.1375,
+ "step": 808
+ },
+ {
+ "epoch": 20.38,
+ "learning_rate": 0.00017588141025641027,
+ "loss": 0.1457,
+ "step": 809
+ },
+ {
+ "epoch": 20.41,
+ "learning_rate": 0.00017548076923076922,
+ "loss": 0.1447,
+ "step": 810
+ },
+ {
+ "epoch": 20.43,
+ "learning_rate": 0.0001750801282051282,
+ "loss": 0.1381,
+ "step": 811
+ },
+ {
+ "epoch": 20.46,
+ "learning_rate": 0.00017467948717948717,
+ "loss": 0.1443,
+ "step": 812
+ },
+ {
+ "epoch": 20.49,
+ "learning_rate": 0.00017427884615384614,
+ "loss": 0.1427,
+ "step": 813
+ },
+ {
+ "epoch": 20.51,
+ "learning_rate": 0.00017387820512820512,
+ "loss": 0.1425,
+ "step": 814
+ },
+ {
+ "epoch": 20.54,
+ "learning_rate": 0.0001734775641025641,
+ "loss": 0.1428,
+ "step": 815
+ },
+ {
+ "epoch": 20.56,
+ "learning_rate": 0.00017307692307692307,
+ "loss": 0.1454,
+ "step": 816
+ },
+ {
+ "epoch": 20.59,
+ "learning_rate": 0.00017267628205128204,
+ "loss": 0.1422,
+ "step": 817
+ },
+ {
+ "epoch": 20.61,
+ "learning_rate": 0.00017227564102564102,
+ "loss": 0.1426,
+ "step": 818
+ },
+ {
+ "epoch": 20.64,
+ "learning_rate": 0.000171875,
+ "loss": 0.1453,
+ "step": 819
+ },
+ {
+ "epoch": 20.66,
+ "learning_rate": 0.000171474358974359,
+ "loss": 0.1472,
+ "step": 820
+ },
+ {
+ "epoch": 20.69,
+ "learning_rate": 0.00017107371794871797,
+ "loss": 0.1479,
+ "step": 821
+ },
+ {
+ "epoch": 20.71,
+ "learning_rate": 0.00017067307692307694,
+ "loss": 0.1387,
+ "step": 822
+ },
+ {
+ "epoch": 20.74,
+ "learning_rate": 0.00017027243589743592,
+ "loss": 0.15,
+ "step": 823
+ },
+ {
+ "epoch": 20.76,
+ "learning_rate": 0.0001698717948717949,
+ "loss": 0.1429,
+ "step": 824
+ },
+ {
+ "epoch": 20.79,
+ "learning_rate": 0.00016947115384615387,
+ "loss": 0.1532,
+ "step": 825
+ },
+ {
+ "epoch": 20.81,
+ "learning_rate": 0.00016907051282051284,
+ "loss": 0.1452,
+ "step": 826
+ },
+ {
+ "epoch": 20.84,
+ "learning_rate": 0.00016866987179487181,
+ "loss": 0.1445,
+ "step": 827
+ },
+ {
+ "epoch": 20.86,
+ "learning_rate": 0.0001682692307692308,
+ "loss": 0.1461,
+ "step": 828
+ },
+ {
+ "epoch": 20.89,
+ "learning_rate": 0.00016786858974358974,
+ "loss": 0.1544,
+ "step": 829
+ },
+ {
+ "epoch": 20.91,
+ "learning_rate": 0.0001674679487179487,
+ "loss": 0.1476,
+ "step": 830
+ },
+ {
+ "epoch": 20.94,
+ "learning_rate": 0.00016706730769230769,
+ "loss": 0.1535,
+ "step": 831
+ },
+ {
+ "epoch": 20.96,
+ "learning_rate": 0.00016666666666666666,
+ "loss": 0.1498,
+ "step": 832
+ },
+ {
+ "epoch": 20.99,
+ "learning_rate": 0.00016626602564102564,
+ "loss": 0.1514,
+ "step": 833
+ },
+ {
+ "epoch": 21.01,
+ "learning_rate": 0.0001658653846153846,
+ "loss": 0.1405,
+ "step": 834
+ },
+ {
+ "epoch": 21.04,
+ "learning_rate": 0.00016546474358974359,
+ "loss": 0.1218,
+ "step": 835
+ },
+ {
+ "epoch": 21.06,
+ "learning_rate": 0.00016506410256410256,
+ "loss": 0.1228,
+ "step": 836
+ },
+ {
+ "epoch": 21.09,
+ "learning_rate": 0.00016466346153846153,
+ "loss": 0.119,
+ "step": 837
+ },
+ {
+ "epoch": 21.11,
+ "learning_rate": 0.0001642628205128205,
+ "loss": 0.1278,
+ "step": 838
+ },
+ {
+ "epoch": 21.14,
+ "learning_rate": 0.00016386217948717948,
+ "loss": 0.1205,
+ "step": 839
+ },
+ {
+ "epoch": 21.17,
+ "learning_rate": 0.00016346153846153846,
+ "loss": 0.1234,
+ "step": 840
+ },
+ {
+ "epoch": 21.19,
+ "learning_rate": 0.00016306089743589743,
+ "loss": 0.1226,
+ "step": 841
+ },
+ {
+ "epoch": 21.22,
+ "learning_rate": 0.0001626602564102564,
+ "loss": 0.1216,
+ "step": 842
+ },
+ {
+ "epoch": 21.24,
+ "learning_rate": 0.00016225961538461538,
+ "loss": 0.1221,
+ "step": 843
+ },
+ {
+ "epoch": 21.27,
+ "learning_rate": 0.00016185897435897436,
+ "loss": 0.1234,
+ "step": 844
+ },
+ {
+ "epoch": 21.29,
+ "learning_rate": 0.00016145833333333336,
+ "loss": 0.1226,
+ "step": 845
+ },
+ {
+ "epoch": 21.32,
+ "learning_rate": 0.00016105769230769233,
+ "loss": 0.1224,
+ "step": 846
+ },
+ {
+ "epoch": 21.34,
+ "learning_rate": 0.0001606570512820513,
+ "loss": 0.1266,
+ "step": 847
+ },
+ {
+ "epoch": 21.37,
+ "learning_rate": 0.00016025641025641028,
+ "loss": 0.1267,
+ "step": 848
+ },
+ {
+ "epoch": 21.39,
+ "learning_rate": 0.00015985576923076923,
+ "loss": 0.129,
+ "step": 849
+ },
+ {
+ "epoch": 21.42,
+ "learning_rate": 0.0001594551282051282,
+ "loss": 0.1213,
+ "step": 850
+ },
+ {
+ "epoch": 21.44,
+ "learning_rate": 0.00015905448717948718,
+ "loss": 0.1273,
+ "step": 851
+ },
+ {
+ "epoch": 21.47,
+ "learning_rate": 0.00015865384615384616,
+ "loss": 0.1352,
+ "step": 852
+ },
+ {
+ "epoch": 21.49,
+ "learning_rate": 0.00015825320512820513,
+ "loss": 0.1328,
+ "step": 853
+ },
+ {
+ "epoch": 21.52,
+ "learning_rate": 0.0001578525641025641,
+ "loss": 0.1253,
+ "step": 854
+ },
+ {
+ "epoch": 21.54,
+ "learning_rate": 0.00015745192307692308,
+ "loss": 0.1267,
+ "step": 855
+ },
+ {
+ "epoch": 21.57,
+ "learning_rate": 0.00015705128205128205,
+ "loss": 0.1325,
+ "step": 856
+ },
+ {
+ "epoch": 21.59,
+ "learning_rate": 0.00015665064102564103,
+ "loss": 0.1267,
+ "step": 857
+ },
+ {
+ "epoch": 21.62,
+ "learning_rate": 0.00015625,
+ "loss": 0.1326,
+ "step": 858
+ },
+ {
+ "epoch": 21.64,
+ "learning_rate": 0.00015584935897435898,
+ "loss": 0.1319,
+ "step": 859
+ },
+ {
+ "epoch": 21.67,
+ "learning_rate": 0.00015544871794871795,
+ "loss": 0.1262,
+ "step": 860
+ },
+ {
+ "epoch": 21.69,
+ "learning_rate": 0.00015504807692307693,
+ "loss": 0.1334,
+ "step": 861
+ },
+ {
+ "epoch": 21.72,
+ "learning_rate": 0.0001546474358974359,
+ "loss": 0.1347,
+ "step": 862
+ },
+ {
+ "epoch": 21.74,
+ "learning_rate": 0.00015424679487179488,
+ "loss": 0.1322,
+ "step": 863
+ },
+ {
+ "epoch": 21.77,
+ "learning_rate": 0.00015384615384615385,
+ "loss": 0.1321,
+ "step": 864
+ },
+ {
+ "epoch": 21.8,
+ "learning_rate": 0.00015344551282051283,
+ "loss": 0.128,
+ "step": 865
+ },
+ {
+ "epoch": 21.82,
+ "learning_rate": 0.0001530448717948718,
+ "loss": 0.1365,
+ "step": 866
+ },
+ {
+ "epoch": 21.85,
+ "learning_rate": 0.00015264423076923078,
+ "loss": 0.137,
+ "step": 867
+ },
+ {
+ "epoch": 21.87,
+ "learning_rate": 0.00015224358974358972,
+ "loss": 0.1322,
+ "step": 868
+ },
+ {
+ "epoch": 21.9,
+ "learning_rate": 0.0001518429487179487,
+ "loss": 0.1341,
+ "step": 869
+ },
+ {
+ "epoch": 21.92,
+ "learning_rate": 0.00015144230769230767,
+ "loss": 0.1334,
+ "step": 870
+ },
+ {
+ "epoch": 21.95,
+ "learning_rate": 0.00015104166666666665,
+ "loss": 0.1339,
+ "step": 871
+ },
+ {
+ "epoch": 21.97,
+ "learning_rate": 0.00015064102564102565,
+ "loss": 0.1301,
+ "step": 872
+ },
+ {
+ "epoch": 22.0,
+ "learning_rate": 0.00015024038461538462,
+ "loss": 0.1325,
+ "step": 873
+ },
+ {
+ "epoch": 22.02,
+ "learning_rate": 0.0001498397435897436,
+ "loss": 0.1153,
+ "step": 874
+ },
+ {
+ "epoch": 22.05,
+ "learning_rate": 0.00014943910256410257,
+ "loss": 0.1113,
+ "step": 875
+ },
+ {
+ "epoch": 22.07,
+ "learning_rate": 0.00014903846153846155,
+ "loss": 0.1138,
+ "step": 876
+ },
+ {
+ "epoch": 22.1,
+ "learning_rate": 0.00014863782051282052,
+ "loss": 0.1136,
+ "step": 877
+ },
+ {
+ "epoch": 22.12,
+ "learning_rate": 0.0001482371794871795,
+ "loss": 0.1096,
+ "step": 878
+ },
+ {
+ "epoch": 22.15,
+ "learning_rate": 0.00014783653846153847,
+ "loss": 0.116,
+ "step": 879
+ },
+ {
+ "epoch": 22.17,
+ "learning_rate": 0.00014743589743589745,
+ "loss": 0.1122,
+ "step": 880
+ },
+ {
+ "epoch": 22.2,
+ "learning_rate": 0.00014703525641025642,
+ "loss": 0.1132,
+ "step": 881
+ },
+ {
+ "epoch": 22.22,
+ "learning_rate": 0.0001466346153846154,
+ "loss": 0.1128,
+ "step": 882
+ },
+ {
+ "epoch": 22.25,
+ "learning_rate": 0.00014623397435897437,
+ "loss": 0.1134,
+ "step": 883
+ },
+ {
+ "epoch": 22.27,
+ "learning_rate": 0.00014583333333333335,
+ "loss": 0.1065,
+ "step": 884
+ },
+ {
+ "epoch": 22.3,
+ "learning_rate": 0.00014543269230769232,
+ "loss": 0.117,
+ "step": 885
+ },
+ {
+ "epoch": 22.32,
+ "learning_rate": 0.0001450320512820513,
+ "loss": 0.112,
+ "step": 886
+ },
+ {
+ "epoch": 22.35,
+ "learning_rate": 0.00014463141025641027,
+ "loss": 0.1145,
+ "step": 887
+ },
+ {
+ "epoch": 22.37,
+ "learning_rate": 0.00014423076923076922,
+ "loss": 0.1128,
+ "step": 888
+ },
+ {
+ "epoch": 22.4,
+ "learning_rate": 0.0001438301282051282,
+ "loss": 0.1132,
+ "step": 889
+ },
+ {
+ "epoch": 22.43,
+ "learning_rate": 0.00014342948717948717,
+ "loss": 0.1143,
+ "step": 890
+ },
+ {
+ "epoch": 22.45,
+ "learning_rate": 0.00014302884615384614,
+ "loss": 0.1172,
+ "step": 891
+ },
+ {
+ "epoch": 22.48,
+ "learning_rate": 0.00014262820512820512,
+ "loss": 0.1143,
+ "step": 892
+ },
+ {
+ "epoch": 22.5,
+ "learning_rate": 0.0001422275641025641,
+ "loss": 0.1164,
+ "step": 893
+ },
+ {
+ "epoch": 22.53,
+ "learning_rate": 0.00014182692307692307,
+ "loss": 0.1197,
+ "step": 894
+ },
+ {
+ "epoch": 22.55,
+ "learning_rate": 0.00014142628205128204,
+ "loss": 0.1152,
+ "step": 895
+ },
+ {
+ "epoch": 22.58,
+ "learning_rate": 0.00014102564102564101,
+ "loss": 0.1243,
+ "step": 896
+ },
+ {
+ "epoch": 22.6,
+ "learning_rate": 0.00014062500000000002,
+ "loss": 0.1167,
+ "step": 897
+ },
+ {
+ "epoch": 22.63,
+ "learning_rate": 0.000140224358974359,
+ "loss": 0.1224,
+ "step": 898
+ },
+ {
+ "epoch": 22.65,
+ "learning_rate": 0.00013982371794871797,
+ "loss": 0.1192,
+ "step": 899
+ },
+ {
+ "epoch": 22.68,
+ "learning_rate": 0.00013942307692307694,
+ "loss": 0.1184,
+ "step": 900
+ },
+ {
+ "epoch": 22.7,
+ "learning_rate": 0.00013902243589743592,
+ "loss": 0.1169,
+ "step": 901
+ },
+ {
+ "epoch": 22.73,
+ "learning_rate": 0.0001386217948717949,
+ "loss": 0.1169,
+ "step": 902
+ },
+ {
+ "epoch": 22.75,
+ "learning_rate": 0.00013822115384615386,
+ "loss": 0.1177,
+ "step": 903
+ },
+ {
+ "epoch": 22.78,
+ "learning_rate": 0.00013782051282051284,
+ "loss": 0.1175,
+ "step": 904
+ },
+ {
+ "epoch": 22.8,
+ "learning_rate": 0.00013741987179487181,
+ "loss": 0.1164,
+ "step": 905
+ },
+ {
+ "epoch": 22.83,
+ "learning_rate": 0.0001370192307692308,
+ "loss": 0.1215,
+ "step": 906
+ },
+ {
+ "epoch": 22.85,
+ "learning_rate": 0.00013661858974358974,
+ "loss": 0.1204,
+ "step": 907
+ },
+ {
+ "epoch": 22.88,
+ "learning_rate": 0.0001362179487179487,
+ "loss": 0.1208,
+ "step": 908
+ },
+ {
+ "epoch": 22.9,
+ "learning_rate": 0.00013581730769230769,
+ "loss": 0.1175,
+ "step": 909
+ },
+ {
+ "epoch": 22.93,
+ "learning_rate": 0.00013541666666666666,
+ "loss": 0.1214,
+ "step": 910
+ },
+ {
+ "epoch": 22.95,
+ "learning_rate": 0.00013501602564102564,
+ "loss": 0.1243,
+ "step": 911
+ },
+ {
+ "epoch": 22.98,
+ "learning_rate": 0.0001346153846153846,
+ "loss": 0.1266,
+ "step": 912
+ },
+ {
+ "epoch": 23.0,
+ "learning_rate": 0.00013421474358974358,
+ "loss": 0.1226,
+ "step": 913
+ },
+ {
+ "epoch": 23.03,
+ "learning_rate": 0.00013381410256410256,
+ "loss": 0.1058,
+ "step": 914
+ },
+ {
+ "epoch": 23.06,
+ "learning_rate": 0.00013341346153846153,
+ "loss": 0.1034,
+ "step": 915
+ },
+ {
+ "epoch": 23.08,
+ "learning_rate": 0.0001330128205128205,
+ "loss": 0.1026,
+ "step": 916
+ },
+ {
+ "epoch": 23.11,
+ "learning_rate": 0.00013261217948717948,
+ "loss": 0.1026,
+ "step": 917
+ },
+ {
+ "epoch": 23.13,
+ "learning_rate": 0.00013221153846153846,
+ "loss": 0.1028,
+ "step": 918
+ },
+ {
+ "epoch": 23.16,
+ "learning_rate": 0.00013181089743589743,
+ "loss": 0.1024,
+ "step": 919
+ },
+ {
+ "epoch": 23.18,
+ "learning_rate": 0.0001314102564102564,
+ "loss": 0.1042,
+ "step": 920
+ },
+ {
+ "epoch": 23.21,
+ "learning_rate": 0.00013100961538461538,
+ "loss": 0.0992,
+ "step": 921
+ },
+ {
+ "epoch": 23.23,
+ "learning_rate": 0.00013060897435897436,
+ "loss": 0.1048,
+ "step": 922
+ },
+ {
+ "epoch": 23.26,
+ "learning_rate": 0.00013020833333333336,
+ "loss": 0.1006,
+ "step": 923
+ },
+ {
+ "epoch": 23.28,
+ "learning_rate": 0.00012980769230769233,
+ "loss": 0.1041,
+ "step": 924
+ },
+ {
+ "epoch": 23.31,
+ "learning_rate": 0.0001294070512820513,
+ "loss": 0.1036,
+ "step": 925
+ },
+ {
+ "epoch": 23.33,
+ "learning_rate": 0.00012900641025641028,
+ "loss": 0.1045,
+ "step": 926
+ },
+ {
+ "epoch": 23.36,
+ "learning_rate": 0.00012860576923076923,
+ "loss": 0.1048,
+ "step": 927
+ },
+ {
+ "epoch": 23.38,
+ "learning_rate": 0.0001282051282051282,
+ "loss": 0.1082,
+ "step": 928
+ },
+ {
+ "epoch": 23.41,
+ "learning_rate": 0.00012780448717948718,
+ "loss": 0.1092,
+ "step": 929
+ },
+ {
+ "epoch": 23.43,
+ "learning_rate": 0.00012740384615384615,
+ "loss": 0.1085,
+ "step": 930
+ },
+ {
+ "epoch": 23.46,
+ "learning_rate": 0.00012700320512820513,
+ "loss": 0.1037,
+ "step": 931
+ },
+ {
+ "epoch": 23.48,
+ "learning_rate": 0.0001266025641025641,
+ "loss": 0.109,
+ "step": 932
+ },
+ {
+ "epoch": 23.51,
+ "learning_rate": 0.00012620192307692308,
+ "loss": 0.1076,
+ "step": 933
+ },
+ {
+ "epoch": 23.53,
+ "learning_rate": 0.00012580128205128205,
+ "loss": 0.1062,
+ "step": 934
+ },
+ {
+ "epoch": 23.56,
+ "learning_rate": 0.00012540064102564103,
+ "loss": 0.1049,
+ "step": 935
+ },
+ {
+ "epoch": 23.58,
+ "learning_rate": 0.000125,
+ "loss": 0.1057,
+ "step": 936
+ },
+ {
+ "epoch": 23.61,
+ "learning_rate": 0.00012459935897435898,
+ "loss": 0.1106,
+ "step": 937
+ },
+ {
+ "epoch": 23.63,
+ "learning_rate": 0.00012419871794871795,
+ "loss": 0.1134,
+ "step": 938
+ },
+ {
+ "epoch": 23.66,
+ "learning_rate": 0.00012379807692307693,
+ "loss": 0.1124,
+ "step": 939
+ },
+ {
+ "epoch": 23.69,
+ "learning_rate": 0.0001233974358974359,
+ "loss": 0.105,
+ "step": 940
+ },
+ {
+ "epoch": 23.71,
+ "learning_rate": 0.00012299679487179488,
+ "loss": 0.113,
+ "step": 941
+ },
+ {
+ "epoch": 23.74,
+ "learning_rate": 0.00012259615384615385,
+ "loss": 0.1104,
+ "step": 942
+ },
+ {
+ "epoch": 23.76,
+ "learning_rate": 0.00012219551282051283,
+ "loss": 0.1136,
+ "step": 943
+ },
+ {
+ "epoch": 23.79,
+ "learning_rate": 0.0001217948717948718,
+ "loss": 0.1104,
+ "step": 944
+ },
+ {
+ "epoch": 23.81,
+ "learning_rate": 0.00012139423076923077,
+ "loss": 0.1054,
+ "step": 945
+ },
+ {
+ "epoch": 23.84,
+ "learning_rate": 0.00012099358974358975,
+ "loss": 0.1109,
+ "step": 946
+ },
+ {
+ "epoch": 23.86,
+ "learning_rate": 0.00012059294871794872,
+ "loss": 0.1062,
+ "step": 947
+ },
+ {
+ "epoch": 23.89,
+ "learning_rate": 0.0001201923076923077,
+ "loss": 0.1185,
+ "step": 948
+ },
+ {
+ "epoch": 23.91,
+ "learning_rate": 0.00011979166666666667,
+ "loss": 0.1138,
+ "step": 949
+ },
+ {
+ "epoch": 23.94,
+ "learning_rate": 0.00011939102564102565,
+ "loss": 0.1119,
+ "step": 950
+ },
+ {
+ "epoch": 23.96,
+ "learning_rate": 0.00011899038461538461,
+ "loss": 0.1114,
+ "step": 951
+ },
+ {
+ "epoch": 23.99,
+ "learning_rate": 0.00011858974358974358,
+ "loss": 0.1122,
+ "step": 952
+ },
+ {
+ "epoch": 24.01,
+ "learning_rate": 0.00011818910256410256,
+ "loss": 0.1051,
+ "step": 953
+ },
+ {
+ "epoch": 24.04,
+ "learning_rate": 0.00011778846153846153,
+ "loss": 0.0924,
+ "step": 954
+ },
+ {
+ "epoch": 24.06,
+ "learning_rate": 0.00011738782051282051,
+ "loss": 0.0977,
+ "step": 955
+ },
+ {
+ "epoch": 24.09,
+ "learning_rate": 0.0001169871794871795,
+ "loss": 0.0945,
+ "step": 956
+ },
+ {
+ "epoch": 24.11,
+ "learning_rate": 0.00011658653846153847,
+ "loss": 0.0958,
+ "step": 957
+ },
+ {
+ "epoch": 24.14,
+ "learning_rate": 0.00011618589743589745,
+ "loss": 0.0934,
+ "step": 958
+ },
+ {
+ "epoch": 24.16,
+ "learning_rate": 0.00011578525641025642,
+ "loss": 0.0952,
+ "step": 959
+ },
+ {
+ "epoch": 24.19,
+ "learning_rate": 0.0001153846153846154,
+ "loss": 0.0993,
+ "step": 960
+ },
+ {
+ "epoch": 24.21,
+ "learning_rate": 0.00011498397435897436,
+ "loss": 0.0978,
+ "step": 961
+ },
+ {
+ "epoch": 24.24,
+ "learning_rate": 0.00011458333333333333,
+ "loss": 0.0984,
+ "step": 962
+ },
+ {
+ "epoch": 24.26,
+ "learning_rate": 0.0001141826923076923,
+ "loss": 0.1019,
+ "step": 963
+ },
+ {
+ "epoch": 24.29,
+ "learning_rate": 0.00011378205128205128,
+ "loss": 0.0994,
+ "step": 964
+ },
+ {
+ "epoch": 24.31,
+ "learning_rate": 0.00011338141025641026,
+ "loss": 0.1006,
+ "step": 965
+ },
+ {
+ "epoch": 24.34,
+ "learning_rate": 0.00011298076923076923,
+ "loss": 0.0971,
+ "step": 966
+ },
+ {
+ "epoch": 24.37,
+ "learning_rate": 0.0001125801282051282,
+ "loss": 0.1008,
+ "step": 967
+ },
+ {
+ "epoch": 24.39,
+ "learning_rate": 0.00011217948717948718,
+ "loss": 0.0948,
+ "step": 968
+ },
+ {
+ "epoch": 24.42,
+ "learning_rate": 0.00011177884615384617,
+ "loss": 0.1026,
+ "step": 969
+ },
+ {
+ "epoch": 24.44,
+ "learning_rate": 0.00011137820512820514,
+ "loss": 0.0966,
+ "step": 970
+ },
+ {
+ "epoch": 24.47,
+ "learning_rate": 0.0001109775641025641,
+ "loss": 0.1004,
+ "step": 971
+ },
+ {
+ "epoch": 24.49,
+ "learning_rate": 0.00011057692307692308,
+ "loss": 0.1031,
+ "step": 972
+ },
+ {
+ "epoch": 24.52,
+ "learning_rate": 0.00011017628205128205,
+ "loss": 0.0984,
+ "step": 973
+ },
+ {
+ "epoch": 24.54,
+ "learning_rate": 0.00010977564102564103,
+ "loss": 0.0988,
+ "step": 974
+ },
+ {
+ "epoch": 24.57,
+ "learning_rate": 0.000109375,
+ "loss": 0.1048,
+ "step": 975
+ },
+ {
+ "epoch": 24.59,
+ "learning_rate": 0.00010897435897435898,
+ "loss": 0.0987,
+ "step": 976
+ },
+ {
+ "epoch": 24.62,
+ "learning_rate": 0.00010857371794871795,
+ "loss": 0.1026,
+ "step": 977
+ },
+ {
+ "epoch": 24.64,
+ "learning_rate": 0.00010817307692307693,
+ "loss": 0.1015,
+ "step": 978
+ },
+ {
+ "epoch": 24.67,
+ "learning_rate": 0.0001077724358974359,
+ "loss": 0.102,
+ "step": 979
+ },
+ {
+ "epoch": 24.69,
+ "learning_rate": 0.00010737179487179486,
+ "loss": 0.1005,
+ "step": 980
+ },
+ {
+ "epoch": 24.72,
+ "learning_rate": 0.00010697115384615384,
+ "loss": 0.1042,
+ "step": 981
+ },
+ {
+ "epoch": 24.74,
+ "learning_rate": 0.00010657051282051283,
+ "loss": 0.1046,
+ "step": 982
+ },
+ {
+ "epoch": 24.77,
+ "learning_rate": 0.0001061698717948718,
+ "loss": 0.0976,
+ "step": 983
+ },
+ {
+ "epoch": 24.79,
+ "learning_rate": 0.00010576923076923077,
+ "loss": 0.1014,
+ "step": 984
+ },
+ {
+ "epoch": 24.82,
+ "learning_rate": 0.00010536858974358975,
+ "loss": 0.1028,
+ "step": 985
+ },
+ {
+ "epoch": 24.84,
+ "learning_rate": 0.00010496794871794872,
+ "loss": 0.1022,
+ "step": 986
+ },
+ {
+ "epoch": 24.87,
+ "learning_rate": 0.0001045673076923077,
+ "loss": 0.1021,
+ "step": 987
+ },
+ {
+ "epoch": 24.89,
+ "learning_rate": 0.00010416666666666667,
+ "loss": 0.1047,
+ "step": 988
+ },
+ {
+ "epoch": 24.92,
+ "learning_rate": 0.00010376602564102565,
+ "loss": 0.1062,
+ "step": 989
+ },
+ {
+ "epoch": 24.94,
+ "learning_rate": 0.00010336538461538461,
+ "loss": 0.1028,
+ "step": 990
+ },
+ {
+ "epoch": 24.97,
+ "learning_rate": 0.00010296474358974358,
+ "loss": 0.1051,
+ "step": 991
+ },
+ {
+ "epoch": 25.0,
+ "learning_rate": 0.00010256410256410256,
+ "loss": 0.1047,
+ "step": 992
+ },
+ {
+ "epoch": 25.02,
+ "learning_rate": 0.00010216346153846153,
+ "loss": 0.097,
+ "step": 993
+ },
+ {
+ "epoch": 25.05,
+ "learning_rate": 0.00010176282051282051,
+ "loss": 0.0963,
+ "step": 994
+ },
+ {
+ "epoch": 25.07,
+ "learning_rate": 0.0001013621794871795,
+ "loss": 0.0888,
+ "step": 995
+ },
+ {
+ "epoch": 25.1,
+ "learning_rate": 0.00010096153846153847,
+ "loss": 0.089,
+ "step": 996
+ },
+ {
+ "epoch": 25.12,
+ "learning_rate": 0.00010056089743589745,
+ "loss": 0.0911,
+ "step": 997
+ },
+ {
+ "epoch": 25.15,
+ "learning_rate": 0.00010016025641025642,
+ "loss": 0.0916,
+ "step": 998
+ },
+ {
+ "epoch": 25.17,
+ "learning_rate": 9.97596153846154e-05,
+ "loss": 0.089,
+ "step": 999
+ },
+ {
+ "epoch": 25.2,
+ "learning_rate": 9.935897435897436e-05,
+ "loss": 0.0883,
+ "step": 1000
+ },
+ {
+ "epoch": 25.22,
+ "learning_rate": 9.895833333333333e-05,
+ "loss": 0.0999,
+ "step": 1001
+ },
+ {
+ "epoch": 25.25,
+ "learning_rate": 9.85576923076923e-05,
+ "loss": 0.0918,
+ "step": 1002
+ },
+ {
+ "epoch": 25.27,
+ "learning_rate": 9.815705128205128e-05,
+ "loss": 0.0953,
+ "step": 1003
+ },
+ {
+ "epoch": 25.3,
+ "learning_rate": 9.775641025641025e-05,
+ "loss": 0.0917,
+ "step": 1004
+ },
+ {
+ "epoch": 25.32,
+ "learning_rate": 9.735576923076923e-05,
+ "loss": 0.0913,
+ "step": 1005
+ },
+ {
+ "epoch": 25.35,
+ "learning_rate": 9.69551282051282e-05,
+ "loss": 0.0912,
+ "step": 1006
+ },
+ {
+ "epoch": 25.37,
+ "learning_rate": 9.655448717948718e-05,
+ "loss": 0.0897,
+ "step": 1007
+ },
+ {
+ "epoch": 25.4,
+ "learning_rate": 9.615384615384617e-05,
+ "loss": 0.0924,
+ "step": 1008
+ },
+ {
+ "epoch": 25.42,
+ "learning_rate": 9.575320512820514e-05,
+ "loss": 0.0939,
+ "step": 1009
+ },
+ {
+ "epoch": 25.45,
+ "learning_rate": 9.53525641025641e-05,
+ "loss": 0.0931,
+ "step": 1010
+ },
+ {
+ "epoch": 25.47,
+ "learning_rate": 9.495192307692308e-05,
+ "loss": 0.0919,
+ "step": 1011
+ },
+ {
+ "epoch": 25.5,
+ "learning_rate": 9.455128205128205e-05,
+ "loss": 0.0926,
+ "step": 1012
+ },
+ {
+ "epoch": 25.52,
+ "learning_rate": 9.415064102564103e-05,
+ "loss": 0.0958,
+ "step": 1013
+ },
+ {
+ "epoch": 25.55,
+ "learning_rate": 9.375e-05,
+ "loss": 0.097,
+ "step": 1014
+ },
+ {
+ "epoch": 25.57,
+ "learning_rate": 9.334935897435898e-05,
+ "loss": 0.0946,
+ "step": 1015
+ },
+ {
+ "epoch": 25.6,
+ "learning_rate": 9.294871794871795e-05,
+ "loss": 0.0943,
+ "step": 1016
+ },
+ {
+ "epoch": 25.63,
+ "learning_rate": 9.254807692307693e-05,
+ "loss": 0.0931,
+ "step": 1017
+ },
+ {
+ "epoch": 25.65,
+ "learning_rate": 9.21474358974359e-05,
+ "loss": 0.094,
+ "step": 1018
+ },
+ {
+ "epoch": 25.68,
+ "learning_rate": 9.174679487179486e-05,
+ "loss": 0.0971,
+ "step": 1019
+ },
+ {
+ "epoch": 25.7,
+ "learning_rate": 9.134615384615384e-05,
+ "loss": 0.0972,
+ "step": 1020
+ },
+ {
+ "epoch": 25.73,
+ "learning_rate": 9.094551282051282e-05,
+ "loss": 0.1061,
+ "step": 1021
+ },
+ {
+ "epoch": 25.75,
+ "learning_rate": 9.05448717948718e-05,
+ "loss": 0.0974,
+ "step": 1022
+ },
+ {
+ "epoch": 25.78,
+ "learning_rate": 9.014423076923077e-05,
+ "loss": 0.0964,
+ "step": 1023
+ },
+ {
+ "epoch": 25.8,
+ "learning_rate": 8.974358974358975e-05,
+ "loss": 0.095,
+ "step": 1024
+ },
+ {
+ "epoch": 25.83,
+ "learning_rate": 8.934294871794872e-05,
+ "loss": 0.1016,
+ "step": 1025
+ },
+ {
+ "epoch": 25.85,
+ "learning_rate": 8.89423076923077e-05,
+ "loss": 0.0959,
+ "step": 1026
+ },
+ {
+ "epoch": 25.88,
+ "learning_rate": 8.854166666666667e-05,
+ "loss": 0.1008,
+ "step": 1027
+ },
+ {
+ "epoch": 25.9,
+ "learning_rate": 8.814102564102565e-05,
+ "loss": 0.1004,
+ "step": 1028
+ },
+ {
+ "epoch": 25.93,
+ "learning_rate": 8.774038461538461e-05,
+ "loss": 0.0959,
+ "step": 1029
+ },
+ {
+ "epoch": 25.95,
+ "learning_rate": 8.733974358974358e-05,
+ "loss": 0.0958,
+ "step": 1030
+ },
+ {
+ "epoch": 25.98,
+ "learning_rate": 8.693910256410256e-05,
+ "loss": 0.0961,
+ "step": 1031
+ },
+ {
+ "epoch": 26.0,
+ "learning_rate": 8.653846153846153e-05,
+ "loss": 0.0957,
+ "step": 1032
+ },
+ {
+ "epoch": 26.03,
+ "learning_rate": 8.613782051282051e-05,
+ "loss": 0.0855,
+ "step": 1033
+ },
+ {
+ "epoch": 26.05,
+ "learning_rate": 8.57371794871795e-05,
+ "loss": 0.0835,
+ "step": 1034
+ },
+ {
+ "epoch": 26.08,
+ "learning_rate": 8.533653846153847e-05,
+ "loss": 0.0886,
+ "step": 1035
+ },
+ {
+ "epoch": 26.1,
+ "learning_rate": 8.493589743589745e-05,
+ "loss": 0.0918,
+ "step": 1036
+ },
+ {
+ "epoch": 26.13,
+ "learning_rate": 8.453525641025642e-05,
+ "loss": 0.0837,
+ "step": 1037
+ },
+ {
+ "epoch": 26.15,
+ "learning_rate": 8.41346153846154e-05,
+ "loss": 0.088,
+ "step": 1038
+ },
+ {
+ "epoch": 26.18,
+ "learning_rate": 8.373397435897436e-05,
+ "loss": 0.0928,
+ "step": 1039
+ },
+ {
+ "epoch": 26.2,
+ "learning_rate": 8.333333333333333e-05,
+ "loss": 0.0871,
+ "step": 1040
+ },
+ {
+ "epoch": 26.23,
+ "learning_rate": 8.29326923076923e-05,
+ "loss": 0.0863,
+ "step": 1041
+ },
+ {
+ "epoch": 26.26,
+ "learning_rate": 8.253205128205128e-05,
+ "loss": 0.0876,
+ "step": 1042
+ },
+ {
+ "epoch": 26.28,
+ "learning_rate": 8.213141025641025e-05,
+ "loss": 0.0888,
+ "step": 1043
+ },
+ {
+ "epoch": 26.31,
+ "learning_rate": 8.173076923076923e-05,
+ "loss": 0.0898,
+ "step": 1044
+ },
+ {
+ "epoch": 26.33,
+ "learning_rate": 8.13301282051282e-05,
+ "loss": 0.09,
+ "step": 1045
+ },
+ {
+ "epoch": 26.36,
+ "learning_rate": 8.092948717948718e-05,
+ "loss": 0.0906,
+ "step": 1046
+ },
+ {
+ "epoch": 26.38,
+ "learning_rate": 8.052884615384617e-05,
+ "loss": 0.091,
+ "step": 1047
+ },
+ {
+ "epoch": 26.41,
+ "learning_rate": 8.012820512820514e-05,
+ "loss": 0.0885,
+ "step": 1048
+ },
+ {
+ "epoch": 26.43,
+ "learning_rate": 7.97275641025641e-05,
+ "loss": 0.0896,
+ "step": 1049
+ },
+ {
+ "epoch": 26.46,
+ "learning_rate": 7.932692307692308e-05,
+ "loss": 0.0891,
+ "step": 1050
+ },
+ {
+ "epoch": 26.48,
+ "learning_rate": 7.892628205128205e-05,
+ "loss": 0.0875,
+ "step": 1051
+ },
+ {
+ "epoch": 26.51,
+ "learning_rate": 7.852564102564103e-05,
+ "loss": 0.0906,
+ "step": 1052
+ },
+ {
+ "epoch": 26.53,
+ "learning_rate": 7.8125e-05,
+ "loss": 0.0913,
+ "step": 1053
+ },
+ {
+ "epoch": 26.56,
+ "learning_rate": 7.772435897435898e-05,
+ "loss": 0.0898,
+ "step": 1054
+ },
+ {
+ "epoch": 26.58,
+ "learning_rate": 7.732371794871795e-05,
+ "loss": 0.0887,
+ "step": 1055
+ },
+ {
+ "epoch": 26.61,
+ "learning_rate": 7.692307692307693e-05,
+ "loss": 0.0916,
+ "step": 1056
+ },
+ {
+ "epoch": 26.63,
+ "learning_rate": 7.65224358974359e-05,
+ "loss": 0.0934,
+ "step": 1057
+ },
+ {
+ "epoch": 26.66,
+ "learning_rate": 7.612179487179486e-05,
+ "loss": 0.0864,
+ "step": 1058
+ },
+ {
+ "epoch": 26.68,
+ "learning_rate": 7.572115384615384e-05,
+ "loss": 0.0929,
+ "step": 1059
+ },
+ {
+ "epoch": 26.71,
+ "learning_rate": 7.532051282051282e-05,
+ "loss": 0.0893,
+ "step": 1060
+ },
+ {
+ "epoch": 26.73,
+ "learning_rate": 7.49198717948718e-05,
+ "loss": 0.0926,
+ "step": 1061
+ },
+ {
+ "epoch": 26.76,
+ "learning_rate": 7.451923076923077e-05,
+ "loss": 0.0931,
+ "step": 1062
+ },
+ {
+ "epoch": 26.78,
+ "learning_rate": 7.411858974358975e-05,
+ "loss": 0.0921,
+ "step": 1063
+ },
+ {
+ "epoch": 26.81,
+ "learning_rate": 7.371794871794872e-05,
+ "loss": 0.0882,
+ "step": 1064
+ },
+ {
+ "epoch": 26.83,
+ "learning_rate": 7.33173076923077e-05,
+ "loss": 0.0937,
+ "step": 1065
+ },
+ {
+ "epoch": 26.86,
+ "learning_rate": 7.291666666666667e-05,
+ "loss": 0.0913,
+ "step": 1066
+ },
+ {
+ "epoch": 26.89,
+ "learning_rate": 7.251602564102565e-05,
+ "loss": 0.0925,
+ "step": 1067
+ },
+ {
+ "epoch": 26.91,
+ "learning_rate": 7.211538461538461e-05,
+ "loss": 0.0942,
+ "step": 1068
+ },
+ {
+ "epoch": 26.94,
+ "learning_rate": 7.171474358974358e-05,
+ "loss": 0.097,
+ "step": 1069
+ },
+ {
+ "epoch": 26.96,
+ "learning_rate": 7.131410256410256e-05,
+ "loss": 0.0922,
+ "step": 1070
+ },
+ {
+ "epoch": 26.99,
+ "learning_rate": 7.091346153846153e-05,
+ "loss": 0.0925,
+ "step": 1071
+ },
+ {
+ "epoch": 27.01,
+ "learning_rate": 7.051282051282051e-05,
+ "loss": 0.0887,
+ "step": 1072
+ },
+ {
+ "epoch": 27.04,
+ "learning_rate": 7.01121794871795e-05,
+ "loss": 0.0831,
+ "step": 1073
+ },
+ {
+ "epoch": 27.06,
+ "learning_rate": 6.971153846153847e-05,
+ "loss": 0.084,
+ "step": 1074
+ },
+ {
+ "epoch": 27.09,
+ "learning_rate": 6.931089743589744e-05,
+ "loss": 0.083,
+ "step": 1075
+ },
+ {
+ "epoch": 27.11,
+ "learning_rate": 6.891025641025642e-05,
+ "loss": 0.0857,
+ "step": 1076
+ },
+ {
+ "epoch": 27.14,
+ "learning_rate": 6.85096153846154e-05,
+ "loss": 0.0837,
+ "step": 1077
+ },
+ {
+ "epoch": 27.16,
+ "learning_rate": 6.810897435897436e-05,
+ "loss": 0.0854,
+ "step": 1078
+ },
+ {
+ "epoch": 27.19,
+ "learning_rate": 6.770833333333333e-05,
+ "loss": 0.0877,
+ "step": 1079
+ },
+ {
+ "epoch": 27.21,
+ "learning_rate": 6.73076923076923e-05,
+ "loss": 0.0861,
+ "step": 1080
+ },
+ {
+ "epoch": 27.24,
+ "learning_rate": 6.690705128205128e-05,
+ "loss": 0.0825,
+ "step": 1081
+ },
+ {
+ "epoch": 27.26,
+ "learning_rate": 6.650641025641025e-05,
+ "loss": 0.0831,
+ "step": 1082
+ },
+ {
+ "epoch": 27.29,
+ "learning_rate": 6.610576923076923e-05,
+ "loss": 0.086,
+ "step": 1083
+ },
+ {
+ "epoch": 27.31,
+ "learning_rate": 6.57051282051282e-05,
+ "loss": 0.0838,
+ "step": 1084
+ },
+ {
+ "epoch": 27.34,
+ "learning_rate": 6.530448717948718e-05,
+ "loss": 0.0857,
+ "step": 1085
+ },
+ {
+ "epoch": 27.36,
+ "learning_rate": 6.490384615384617e-05,
+ "loss": 0.0836,
+ "step": 1086
+ },
+ {
+ "epoch": 27.39,
+ "learning_rate": 6.450320512820514e-05,
+ "loss": 0.0875,
+ "step": 1087
+ },
+ {
+ "epoch": 27.41,
+ "learning_rate": 6.41025641025641e-05,
+ "loss": 0.0848,
+ "step": 1088
+ },
+ {
+ "epoch": 27.44,
+ "learning_rate": 6.370192307692308e-05,
+ "loss": 0.084,
+ "step": 1089
+ },
+ {
+ "epoch": 27.46,
+ "learning_rate": 6.330128205128205e-05,
+ "loss": 0.0845,
+ "step": 1090
+ },
+ {
+ "epoch": 27.49,
+ "learning_rate": 6.290064102564103e-05,
+ "loss": 0.0876,
+ "step": 1091
+ },
+ {
+ "epoch": 27.51,
+ "learning_rate": 6.25e-05,
+ "loss": 0.0907,
+ "step": 1092
+ },
+ {
+ "epoch": 27.54,
+ "learning_rate": 6.209935897435898e-05,
+ "loss": 0.0864,
+ "step": 1093
+ },
+ {
+ "epoch": 27.57,
+ "learning_rate": 6.169871794871795e-05,
+ "loss": 0.087,
+ "step": 1094
+ },
+ {
+ "epoch": 27.59,
+ "learning_rate": 6.129807692307693e-05,
+ "loss": 0.0902,
+ "step": 1095
+ },
+ {
+ "epoch": 27.62,
+ "learning_rate": 6.08974358974359e-05,
+ "loss": 0.0862,
+ "step": 1096
+ },
+ {
+ "epoch": 27.64,
+ "learning_rate": 6.0496794871794875e-05,
+ "loss": 0.0864,
+ "step": 1097
+ },
+ {
+ "epoch": 27.67,
+ "learning_rate": 6.009615384615385e-05,
+ "loss": 0.0868,
+ "step": 1098
+ },
+ {
+ "epoch": 27.69,
+ "learning_rate": 5.9695512820512824e-05,
+ "loss": 0.0905,
+ "step": 1099
+ },
+ {
+ "epoch": 27.72,
+ "learning_rate": 5.929487179487179e-05,
+ "loss": 0.0881,
+ "step": 1100
+ },
+ {
+ "epoch": 27.74,
+ "learning_rate": 5.889423076923077e-05,
+ "loss": 0.0883,
+ "step": 1101
+ },
+ {
+ "epoch": 27.77,
+ "learning_rate": 5.849358974358975e-05,
+ "loss": 0.0852,
+ "step": 1102
+ },
+ {
+ "epoch": 27.79,
+ "learning_rate": 5.809294871794872e-05,
+ "loss": 0.0838,
+ "step": 1103
+ },
+ {
+ "epoch": 27.82,
+ "learning_rate": 5.76923076923077e-05,
+ "loss": 0.0866,
+ "step": 1104
+ },
+ {
+ "epoch": 27.84,
+ "learning_rate": 5.7291666666666666e-05,
+ "loss": 0.0893,
+ "step": 1105
+ },
+ {
+ "epoch": 27.87,
+ "learning_rate": 5.689102564102564e-05,
+ "loss": 0.0881,
+ "step": 1106
+ },
+ {
+ "epoch": 27.89,
+ "learning_rate": 5.6490384615384615e-05,
+ "loss": 0.0889,
+ "step": 1107
+ },
+ {
+ "epoch": 27.92,
+ "learning_rate": 5.608974358974359e-05,
+ "loss": 0.087,
+ "step": 1108
+ },
+ {
+ "epoch": 27.94,
+ "learning_rate": 5.568910256410257e-05,
+ "loss": 0.089,
+ "step": 1109
+ },
+ {
+ "epoch": 27.97,
+ "learning_rate": 5.528846153846154e-05,
+ "loss": 0.0889,
+ "step": 1110
+ },
+ {
+ "epoch": 27.99,
+ "learning_rate": 5.4887820512820514e-05,
+ "loss": 0.09,
+ "step": 1111
+ },
+ {
+ "epoch": 28.02,
+ "learning_rate": 5.448717948717949e-05,
+ "loss": 0.0838,
+ "step": 1112
+ },
+ {
+ "epoch": 28.04,
+ "learning_rate": 5.408653846153846e-05,
+ "loss": 0.0834,
+ "step": 1113
+ },
+ {
+ "epoch": 28.07,
+ "learning_rate": 5.368589743589743e-05,
+ "loss": 0.0831,
+ "step": 1114
+ },
+ {
+ "epoch": 28.09,
+ "learning_rate": 5.328525641025641e-05,
+ "loss": 0.0804,
+ "step": 1115
+ },
+ {
+ "epoch": 28.12,
+ "learning_rate": 5.288461538461539e-05,
+ "loss": 0.0823,
+ "step": 1116
+ },
+ {
+ "epoch": 28.14,
+ "learning_rate": 5.248397435897436e-05,
+ "loss": 0.0823,
+ "step": 1117
+ },
+ {
+ "epoch": 28.17,
+ "learning_rate": 5.208333333333334e-05,
+ "loss": 0.0778,
+ "step": 1118
+ },
+ {
+ "epoch": 28.2,
+ "learning_rate": 5.1682692307692305e-05,
+ "loss": 0.084,
+ "step": 1119
+ },
+ {
+ "epoch": 28.22,
+ "learning_rate": 5.128205128205128e-05,
+ "loss": 0.0816,
+ "step": 1120
+ },
+ {
+ "epoch": 28.25,
+ "learning_rate": 5.0881410256410254e-05,
+ "loss": 0.0827,
+ "step": 1121
+ },
+ {
+ "epoch": 28.27,
+ "learning_rate": 5.0480769230769235e-05,
+ "loss": 0.0842,
+ "step": 1122
+ },
+ {
+ "epoch": 28.3,
+ "learning_rate": 5.008012820512821e-05,
+ "loss": 0.0816,
+ "step": 1123
+ },
+ {
+ "epoch": 28.32,
+ "learning_rate": 4.967948717948718e-05,
+ "loss": 0.0853,
+ "step": 1124
+ },
+ {
+ "epoch": 28.35,
+ "learning_rate": 4.927884615384615e-05,
+ "loss": 0.0793,
+ "step": 1125
+ },
+ {
+ "epoch": 28.37,
+ "learning_rate": 4.887820512820513e-05,
+ "loss": 0.0831,
+ "step": 1126
+ },
+ {
+ "epoch": 28.4,
+ "learning_rate": 4.84775641025641e-05,
+ "loss": 0.0824,
+ "step": 1127
+ },
+ {
+ "epoch": 28.42,
+ "learning_rate": 4.8076923076923084e-05,
+ "loss": 0.0851,
+ "step": 1128
+ },
+ {
+ "epoch": 28.45,
+ "learning_rate": 4.767628205128205e-05,
+ "loss": 0.0842,
+ "step": 1129
+ },
+ {
+ "epoch": 28.47,
+ "learning_rate": 4.7275641025641026e-05,
+ "loss": 0.0806,
+ "step": 1130
+ },
+ {
+ "epoch": 28.5,
+ "learning_rate": 4.6875e-05,
+ "loss": 0.0849,
+ "step": 1131
+ },
+ {
+ "epoch": 28.52,
+ "learning_rate": 4.6474358974358976e-05,
+ "loss": 0.0823,
+ "step": 1132
+ },
+ {
+ "epoch": 28.55,
+ "learning_rate": 4.607371794871795e-05,
+ "loss": 0.0841,
+ "step": 1133
+ },
+ {
+ "epoch": 28.57,
+ "learning_rate": 4.567307692307692e-05,
+ "loss": 0.0842,
+ "step": 1134
+ },
+ {
+ "epoch": 28.6,
+ "learning_rate": 4.52724358974359e-05,
+ "loss": 0.0833,
+ "step": 1135
+ },
+ {
+ "epoch": 28.62,
+ "learning_rate": 4.4871794871794874e-05,
+ "loss": 0.082,
+ "step": 1136
+ },
+ {
+ "epoch": 28.65,
+ "learning_rate": 4.447115384615385e-05,
+ "loss": 0.0867,
+ "step": 1137
+ },
+ {
+ "epoch": 28.67,
+ "learning_rate": 4.4070512820512824e-05,
+ "loss": 0.0845,
+ "step": 1138
+ },
+ {
+ "epoch": 28.7,
+ "learning_rate": 4.366987179487179e-05,
+ "loss": 0.0853,
+ "step": 1139
+ },
+ {
+ "epoch": 28.72,
+ "learning_rate": 4.3269230769230766e-05,
+ "loss": 0.0835,
+ "step": 1140
+ },
+ {
+ "epoch": 28.75,
+ "learning_rate": 4.286858974358975e-05,
+ "loss": 0.0884,
+ "step": 1141
+ },
+ {
+ "epoch": 28.77,
+ "learning_rate": 4.246794871794872e-05,
+ "loss": 0.0857,
+ "step": 1142
+ },
+ {
+ "epoch": 28.8,
+ "learning_rate": 4.20673076923077e-05,
+ "loss": 0.084,
+ "step": 1143
+ },
+ {
+ "epoch": 28.83,
+ "learning_rate": 4.1666666666666665e-05,
+ "loss": 0.0843,
+ "step": 1144
+ },
+ {
+ "epoch": 28.85,
+ "learning_rate": 4.126602564102564e-05,
+ "loss": 0.0827,
+ "step": 1145
+ },
+ {
+ "epoch": 28.88,
+ "learning_rate": 4.0865384615384615e-05,
+ "loss": 0.0826,
+ "step": 1146
+ },
+ {
+ "epoch": 28.9,
+ "learning_rate": 4.046474358974359e-05,
+ "loss": 0.0851,
+ "step": 1147
+ },
+ {
+ "epoch": 28.93,
+ "learning_rate": 4.006410256410257e-05,
+ "loss": 0.0851,
+ "step": 1148
+ },
+ {
+ "epoch": 28.95,
+ "learning_rate": 3.966346153846154e-05,
+ "loss": 0.0848,
+ "step": 1149
+ },
+ {
+ "epoch": 28.98,
+ "learning_rate": 3.9262820512820513e-05,
+ "loss": 0.0828,
+ "step": 1150
+ },
+ {
+ "epoch": 29.0,
+ "learning_rate": 3.886217948717949e-05,
+ "loss": 0.0834,
+ "step": 1151
+ },
+ {
+ "epoch": 29.03,
+ "learning_rate": 3.846153846153846e-05,
+ "loss": 0.0815,
+ "step": 1152
+ },
+ {
+ "epoch": 29.05,
+ "learning_rate": 3.806089743589743e-05,
+ "loss": 0.0806,
+ "step": 1153
+ },
+ {
+ "epoch": 29.08,
+ "learning_rate": 3.766025641025641e-05,
+ "loss": 0.0787,
+ "step": 1154
+ },
+ {
+ "epoch": 29.1,
+ "learning_rate": 3.725961538461539e-05,
+ "loss": 0.0772,
+ "step": 1155
+ },
+ {
+ "epoch": 29.13,
+ "learning_rate": 3.685897435897436e-05,
+ "loss": 0.0794,
+ "step": 1156
+ },
+ {
+ "epoch": 29.15,
+ "learning_rate": 3.6458333333333336e-05,
+ "loss": 0.0799,
+ "step": 1157
+ },
+ {
+ "epoch": 29.18,
+ "learning_rate": 3.6057692307692304e-05,
+ "loss": 0.0803,
+ "step": 1158
+ },
+ {
+ "epoch": 29.2,
+ "learning_rate": 3.565705128205128e-05,
+ "loss": 0.0829,
+ "step": 1159
+ },
+ {
+ "epoch": 29.23,
+ "learning_rate": 3.5256410256410254e-05,
+ "loss": 0.0795,
+ "step": 1160
+ },
+ {
+ "epoch": 29.25,
+ "learning_rate": 3.4855769230769235e-05,
+ "loss": 0.0795,
+ "step": 1161
+ },
+ {
+ "epoch": 29.28,
+ "learning_rate": 3.445512820512821e-05,
+ "loss": 0.0786,
+ "step": 1162
+ },
+ {
+ "epoch": 29.3,
+ "learning_rate": 3.405448717948718e-05,
+ "loss": 0.0778,
+ "step": 1163
+ },
+ {
+ "epoch": 29.33,
+ "learning_rate": 3.365384615384615e-05,
+ "loss": 0.0774,
+ "step": 1164
+ },
+ {
+ "epoch": 29.35,
+ "learning_rate": 3.325320512820513e-05,
+ "loss": 0.0806,
+ "step": 1165
+ },
+ {
+ "epoch": 29.38,
+ "learning_rate": 3.28525641025641e-05,
+ "loss": 0.0789,
+ "step": 1166
+ },
+ {
+ "epoch": 29.4,
+ "learning_rate": 3.245192307692308e-05,
+ "loss": 0.0788,
+ "step": 1167
+ },
+ {
+ "epoch": 29.43,
+ "learning_rate": 3.205128205128205e-05,
+ "loss": 0.0825,
+ "step": 1168
+ },
+ {
+ "epoch": 29.46,
+ "learning_rate": 3.1650641025641026e-05,
+ "loss": 0.0804,
+ "step": 1169
+ },
+ {
+ "epoch": 29.48,
+ "learning_rate": 3.125e-05,
+ "loss": 0.0806,
+ "step": 1170
+ },
+ {
+ "epoch": 29.51,
+ "learning_rate": 3.0849358974358975e-05,
+ "loss": 0.0824,
+ "step": 1171
+ },
+ {
+ "epoch": 29.53,
+ "learning_rate": 3.044871794871795e-05,
+ "loss": 0.0804,
+ "step": 1172
+ },
+ {
+ "epoch": 29.56,
+ "learning_rate": 3.0048076923076925e-05,
+ "loss": 0.0831,
+ "step": 1173
+ },
+ {
+ "epoch": 29.58,
+ "learning_rate": 2.9647435897435896e-05,
+ "loss": 0.0796,
+ "step": 1174
+ },
+ {
+ "epoch": 29.61,
+ "learning_rate": 2.9246794871794874e-05,
+ "loss": 0.0816,
+ "step": 1175
+ },
+ {
+ "epoch": 29.63,
+ "learning_rate": 2.884615384615385e-05,
+ "loss": 0.0862,
+ "step": 1176
+ },
+ {
+ "epoch": 29.66,
+ "learning_rate": 2.844551282051282e-05,
+ "loss": 0.0825,
+ "step": 1177
+ },
+ {
+ "epoch": 29.68,
+ "learning_rate": 2.8044871794871795e-05,
+ "loss": 0.082,
+ "step": 1178
+ },
+ {
+ "epoch": 29.71,
+ "learning_rate": 2.764423076923077e-05,
+ "loss": 0.082,
+ "step": 1179
+ },
+ {
+ "epoch": 29.73,
+ "learning_rate": 2.7243589743589744e-05,
+ "loss": 0.0844,
+ "step": 1180
+ },
+ {
+ "epoch": 29.76,
+ "learning_rate": 2.6842948717948716e-05,
+ "loss": 0.0823,
+ "step": 1181
+ },
+ {
+ "epoch": 29.78,
+ "learning_rate": 2.6442307692307694e-05,
+ "loss": 0.0822,
+ "step": 1182
+ },
+ {
+ "epoch": 29.81,
+ "learning_rate": 2.604166666666667e-05,
+ "loss": 0.0833,
+ "step": 1183
+ },
+ {
+ "epoch": 29.83,
+ "learning_rate": 2.564102564102564e-05,
+ "loss": 0.0853,
+ "step": 1184
+ },
+ {
+ "epoch": 29.86,
+ "learning_rate": 2.5240384615384618e-05,
+ "loss": 0.0803,
+ "step": 1185
+ },
+ {
+ "epoch": 29.88,
+ "learning_rate": 2.483974358974359e-05,
+ "loss": 0.0809,
+ "step": 1186
+ },
+ {
+ "epoch": 29.91,
+ "learning_rate": 2.4439102564102564e-05,
+ "loss": 0.0828,
+ "step": 1187
+ },
+ {
+ "epoch": 29.93,
+ "learning_rate": 2.4038461538461542e-05,
+ "loss": 0.0815,
+ "step": 1188
+ },
+ {
+ "epoch": 29.96,
+ "learning_rate": 2.3637820512820513e-05,
+ "loss": 0.0847,
+ "step": 1189
+ },
+ {
+ "epoch": 29.98,
+ "learning_rate": 2.3237179487179488e-05,
+ "loss": 0.0801,
+ "step": 1190
+ },
+ {
+ "epoch": 30.01,
+ "learning_rate": 2.283653846153846e-05,
+ "loss": 0.0794,
+ "step": 1191
+ },
+ {
+ "epoch": 30.03,
+ "learning_rate": 2.2435897435897437e-05,
+ "loss": 0.0779,
+ "step": 1192
+ },
+ {
+ "epoch": 30.06,
+ "learning_rate": 2.2035256410256412e-05,
+ "loss": 0.0769,
+ "step": 1193
+ },
+ {
+ "epoch": 30.09,
+ "learning_rate": 2.1634615384615383e-05,
+ "loss": 0.0766,
+ "step": 1194
+ },
+ {
+ "epoch": 30.11,
+ "learning_rate": 2.123397435897436e-05,
+ "loss": 0.0778,
+ "step": 1195
+ },
+ {
+ "epoch": 30.14,
+ "learning_rate": 2.0833333333333333e-05,
+ "loss": 0.0811,
+ "step": 1196
+ },
+ {
+ "epoch": 30.16,
+ "learning_rate": 2.0432692307692307e-05,
+ "loss": 0.0779,
+ "step": 1197
+ },
+ {
+ "epoch": 30.19,
+ "learning_rate": 2.0032051282051285e-05,
+ "loss": 0.0788,
+ "step": 1198
+ },
+ {
+ "epoch": 30.21,
+ "learning_rate": 1.9631410256410257e-05,
+ "loss": 0.0805,
+ "step": 1199
+ },
+ {
+ "epoch": 30.24,
+ "learning_rate": 1.923076923076923e-05,
+ "loss": 0.0791,
+ "step": 1200
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 6.758843226530304e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-1200/training_args.bin b/checkpoint-1200/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-1200/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-300/README.md b/checkpoint-300/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-300/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-300/adapter_config.json b/checkpoint-300/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-300/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-300/adapter_model.bin b/checkpoint-300/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..e9c1465e997f91f33c75ca87f9cabf6aeaeac47a
--- /dev/null
+++ b/checkpoint-300/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5dfc1b8458803ffd860d90eb4599c2284b60aa39a418a44a3966aac16492e688
+size 39409357
diff --git a/checkpoint-300/optimizer.pt b/checkpoint-300/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..25207d6f98ea3eb304dd0549907d4050a22b3aa9
--- /dev/null
+++ b/checkpoint-300/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4050c9e817810a65825f67449896912c0286675a222fadfa4cf24a2ea357f5aa
+size 78844421
diff --git a/checkpoint-300/rng_state.pth b/checkpoint-300/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..a0e229d77fa9bb9693355411d7e7ba7367c65223
--- /dev/null
+++ b/checkpoint-300/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:36fc71bd44bd7f04f2599c5dface64c517de1a7ab7bac3600f3f6470c6c72673
+size 14575
diff --git a/checkpoint-300/scheduler.pt b/checkpoint-300/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..83b3ab07ce1c15256928fe8ce2d4749b096a089c
--- /dev/null
+++ b/checkpoint-300/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2173731890e5f0214a48e0accca0e2767cccd021afad070b099ba172c002c0b0
+size 627
diff --git a/checkpoint-300/special_tokens_map.json b/checkpoint-300/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-300/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-300/tokenizer.json b/checkpoint-300/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-300/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-300/tokenizer_config.json b/checkpoint-300/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-300/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-300/trainer_state.json b/checkpoint-300/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..aa46f9eaebbbc0d9b593b7f4e9a11b2ac690e61a
--- /dev/null
+++ b/checkpoint-300/trainer_state.json
@@ -0,0 +1,1819 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 7.559055118110236,
+ "eval_steps": 500,
+ "global_step": 300,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 1.6887156698840064e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-300/training_args.bin b/checkpoint-300/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-300/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-400/README.md b/checkpoint-400/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-400/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-400/adapter_config.json b/checkpoint-400/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-400/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-400/adapter_model.bin b/checkpoint-400/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..7f13cdedec36f680c97e7994548be22b903e8ea9
--- /dev/null
+++ b/checkpoint-400/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:04ef1cda8fd93cc8b65f531e16b8e1d8b4cdbebaab587c409f886214406138b2
+size 39409357
diff --git a/checkpoint-400/optimizer.pt b/checkpoint-400/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d5bbf40cdc5a7a34e7fcd0f37c66ae0a68ea25e4
--- /dev/null
+++ b/checkpoint-400/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ad6ede7bca4b1fee9b3db1aa2d5d7d19ee2339a27761954fd1ec8e2552b307db
+size 78844421
diff --git a/checkpoint-400/rng_state.pth b/checkpoint-400/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..df08b4983b9f5bfcf57aaaceb89bc51479bfd216
--- /dev/null
+++ b/checkpoint-400/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ab62043be50b93d4eb28964be2d945176db3d64fe73ddd052a7656ba9141c683
+size 14575
diff --git a/checkpoint-400/scheduler.pt b/checkpoint-400/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..0b7d9e3816349d7463a0d6101a63282085ff0d61
--- /dev/null
+++ b/checkpoint-400/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5400918926c7e5e51234021a93d05c418b96235f01018a499ecd13536e8888bf
+size 627
diff --git a/checkpoint-400/special_tokens_map.json b/checkpoint-400/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-400/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-400/tokenizer.json b/checkpoint-400/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-400/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-400/tokenizer_config.json b/checkpoint-400/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-400/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-400/trainer_state.json b/checkpoint-400/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..56efbe38b94c1b8981bb1deed21807dd5bf92fe7
--- /dev/null
+++ b/checkpoint-400/trainer_state.json
@@ -0,0 +1,2419 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 10.078740157480315,
+ "eval_steps": 500,
+ "global_step": 400,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 2.2507441328246784e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-400/training_args.bin b/checkpoint-400/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-400/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-500/README.md b/checkpoint-500/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-500/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-500/adapter_config.json b/checkpoint-500/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-500/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-500/adapter_model.bin b/checkpoint-500/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4e949c50b35b26dee2ce6ef624eba77340636cab
--- /dev/null
+++ b/checkpoint-500/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:96e2b1f5229f7d4ae1f40cebcee71ae1174e1f64139fc9416b86fd997c65cfb1
+size 39409357
diff --git a/checkpoint-500/optimizer.pt b/checkpoint-500/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..211d84f0a4583637c6040f56103d478e4a013e3c
--- /dev/null
+++ b/checkpoint-500/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8185cfd65326f0053ee8f7153de24caff5687d41d9cfc8cdb9515e2ef39053e2
+size 78844421
diff --git a/checkpoint-500/rng_state.pth b/checkpoint-500/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..961069ec281e351c01eb6735b97b75dd3d8d33b7
--- /dev/null
+++ b/checkpoint-500/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:696b4cef517e79a03215d5c5fccde0c73d2d962fe70ebab1e472f650136142b2
+size 14575
diff --git a/checkpoint-500/scheduler.pt b/checkpoint-500/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..31e8b00ab047ff86837ffa1ced4a89e09fc63ae2
--- /dev/null
+++ b/checkpoint-500/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0c799ef9044026ca330c8087d78d028c5d503f411779db75cac9194336b3f8c3
+size 627
diff --git a/checkpoint-500/special_tokens_map.json b/checkpoint-500/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-500/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-500/tokenizer.json b/checkpoint-500/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-500/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-500/tokenizer_config.json b/checkpoint-500/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-500/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-500/trainer_state.json b/checkpoint-500/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..f630d21734fbbb08ddd03cabc449b7e8ecba24ec
--- /dev/null
+++ b/checkpoint-500/trainer_state.json
@@ -0,0 +1,3019 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 12.598425196850394,
+ "eval_steps": 500,
+ "global_step": 500,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ },
+ {
+ "epoch": 10.1,
+ "learning_rate": 0.00033934294871794873,
+ "loss": 0.7635,
+ "step": 401
+ },
+ {
+ "epoch": 10.13,
+ "learning_rate": 0.00033894230769230773,
+ "loss": 0.7713,
+ "step": 402
+ },
+ {
+ "epoch": 10.15,
+ "learning_rate": 0.0003385416666666667,
+ "loss": 0.827,
+ "step": 403
+ },
+ {
+ "epoch": 10.18,
+ "learning_rate": 0.0003381410256410257,
+ "loss": 0.7844,
+ "step": 404
+ },
+ {
+ "epoch": 10.2,
+ "learning_rate": 0.00033774038461538463,
+ "loss": 0.7839,
+ "step": 405
+ },
+ {
+ "epoch": 10.23,
+ "learning_rate": 0.00033733974358974363,
+ "loss": 0.7702,
+ "step": 406
+ },
+ {
+ "epoch": 10.26,
+ "learning_rate": 0.0003369391025641026,
+ "loss": 0.7855,
+ "step": 407
+ },
+ {
+ "epoch": 10.28,
+ "learning_rate": 0.0003365384615384616,
+ "loss": 0.7669,
+ "step": 408
+ },
+ {
+ "epoch": 10.31,
+ "learning_rate": 0.0003361378205128205,
+ "loss": 0.8054,
+ "step": 409
+ },
+ {
+ "epoch": 10.33,
+ "learning_rate": 0.0003357371794871795,
+ "loss": 0.7792,
+ "step": 410
+ },
+ {
+ "epoch": 10.36,
+ "learning_rate": 0.0003353365384615384,
+ "loss": 0.7986,
+ "step": 411
+ },
+ {
+ "epoch": 10.38,
+ "learning_rate": 0.0003349358974358974,
+ "loss": 0.8269,
+ "step": 412
+ },
+ {
+ "epoch": 10.41,
+ "learning_rate": 0.00033453525641025637,
+ "loss": 0.7923,
+ "step": 413
+ },
+ {
+ "epoch": 10.43,
+ "learning_rate": 0.00033413461538461537,
+ "loss": 0.8242,
+ "step": 414
+ },
+ {
+ "epoch": 10.46,
+ "learning_rate": 0.0003337339743589743,
+ "loss": 0.8496,
+ "step": 415
+ },
+ {
+ "epoch": 10.48,
+ "learning_rate": 0.0003333333333333333,
+ "loss": 0.8236,
+ "step": 416
+ },
+ {
+ "epoch": 10.51,
+ "learning_rate": 0.0003329326923076923,
+ "loss": 0.8685,
+ "step": 417
+ },
+ {
+ "epoch": 10.53,
+ "learning_rate": 0.00033253205128205127,
+ "loss": 0.8542,
+ "step": 418
+ },
+ {
+ "epoch": 10.56,
+ "learning_rate": 0.0003321314102564103,
+ "loss": 0.8185,
+ "step": 419
+ },
+ {
+ "epoch": 10.58,
+ "learning_rate": 0.0003317307692307692,
+ "loss": 0.8477,
+ "step": 420
+ },
+ {
+ "epoch": 10.61,
+ "learning_rate": 0.0003313301282051282,
+ "loss": 0.8372,
+ "step": 421
+ },
+ {
+ "epoch": 10.63,
+ "learning_rate": 0.00033092948717948717,
+ "loss": 0.8731,
+ "step": 422
+ },
+ {
+ "epoch": 10.66,
+ "learning_rate": 0.00033052884615384617,
+ "loss": 0.8322,
+ "step": 423
+ },
+ {
+ "epoch": 10.68,
+ "learning_rate": 0.0003301282051282051,
+ "loss": 0.8514,
+ "step": 424
+ },
+ {
+ "epoch": 10.71,
+ "learning_rate": 0.0003297275641025641,
+ "loss": 0.8564,
+ "step": 425
+ },
+ {
+ "epoch": 10.73,
+ "learning_rate": 0.00032932692307692307,
+ "loss": 0.8278,
+ "step": 426
+ },
+ {
+ "epoch": 10.76,
+ "learning_rate": 0.00032892628205128207,
+ "loss": 0.8408,
+ "step": 427
+ },
+ {
+ "epoch": 10.78,
+ "learning_rate": 0.000328525641025641,
+ "loss": 0.8438,
+ "step": 428
+ },
+ {
+ "epoch": 10.81,
+ "learning_rate": 0.000328125,
+ "loss": 0.8862,
+ "step": 429
+ },
+ {
+ "epoch": 10.83,
+ "learning_rate": 0.00032772435897435897,
+ "loss": 0.8631,
+ "step": 430
+ },
+ {
+ "epoch": 10.86,
+ "learning_rate": 0.00032732371794871797,
+ "loss": 0.883,
+ "step": 431
+ },
+ {
+ "epoch": 10.89,
+ "learning_rate": 0.0003269230769230769,
+ "loss": 0.8644,
+ "step": 432
+ },
+ {
+ "epoch": 10.91,
+ "learning_rate": 0.0003265224358974359,
+ "loss": 0.8907,
+ "step": 433
+ },
+ {
+ "epoch": 10.94,
+ "learning_rate": 0.00032612179487179487,
+ "loss": 0.9141,
+ "step": 434
+ },
+ {
+ "epoch": 10.96,
+ "learning_rate": 0.00032572115384615387,
+ "loss": 0.8877,
+ "step": 435
+ },
+ {
+ "epoch": 10.99,
+ "learning_rate": 0.0003253205128205128,
+ "loss": 0.8521,
+ "step": 436
+ },
+ {
+ "epoch": 11.01,
+ "learning_rate": 0.0003249198717948718,
+ "loss": 0.7842,
+ "step": 437
+ },
+ {
+ "epoch": 11.04,
+ "learning_rate": 0.00032451923076923077,
+ "loss": 0.6516,
+ "step": 438
+ },
+ {
+ "epoch": 11.06,
+ "learning_rate": 0.00032411858974358977,
+ "loss": 0.6549,
+ "step": 439
+ },
+ {
+ "epoch": 11.09,
+ "learning_rate": 0.0003237179487179487,
+ "loss": 0.6736,
+ "step": 440
+ },
+ {
+ "epoch": 11.11,
+ "learning_rate": 0.0003233173076923077,
+ "loss": 0.6625,
+ "step": 441
+ },
+ {
+ "epoch": 11.14,
+ "learning_rate": 0.0003229166666666667,
+ "loss": 0.6466,
+ "step": 442
+ },
+ {
+ "epoch": 11.16,
+ "learning_rate": 0.00032251602564102567,
+ "loss": 0.6362,
+ "step": 443
+ },
+ {
+ "epoch": 11.19,
+ "learning_rate": 0.00032211538461538467,
+ "loss": 0.674,
+ "step": 444
+ },
+ {
+ "epoch": 11.21,
+ "learning_rate": 0.0003217147435897436,
+ "loss": 0.6509,
+ "step": 445
+ },
+ {
+ "epoch": 11.24,
+ "learning_rate": 0.0003213141025641026,
+ "loss": 0.618,
+ "step": 446
+ },
+ {
+ "epoch": 11.26,
+ "learning_rate": 0.00032091346153846156,
+ "loss": 0.6654,
+ "step": 447
+ },
+ {
+ "epoch": 11.29,
+ "learning_rate": 0.00032051282051282057,
+ "loss": 0.6632,
+ "step": 448
+ },
+ {
+ "epoch": 11.31,
+ "learning_rate": 0.00032011217948717946,
+ "loss": 0.6599,
+ "step": 449
+ },
+ {
+ "epoch": 11.34,
+ "learning_rate": 0.00031971153846153846,
+ "loss": 0.6537,
+ "step": 450
+ },
+ {
+ "epoch": 11.36,
+ "learning_rate": 0.0003193108974358974,
+ "loss": 0.678,
+ "step": 451
+ },
+ {
+ "epoch": 11.39,
+ "learning_rate": 0.0003189102564102564,
+ "loss": 0.684,
+ "step": 452
+ },
+ {
+ "epoch": 11.41,
+ "learning_rate": 0.00031850961538461536,
+ "loss": 0.6419,
+ "step": 453
+ },
+ {
+ "epoch": 11.44,
+ "learning_rate": 0.00031810897435897436,
+ "loss": 0.6614,
+ "step": 454
+ },
+ {
+ "epoch": 11.46,
+ "learning_rate": 0.0003177083333333333,
+ "loss": 0.7037,
+ "step": 455
+ },
+ {
+ "epoch": 11.49,
+ "learning_rate": 0.0003173076923076923,
+ "loss": 0.6944,
+ "step": 456
+ },
+ {
+ "epoch": 11.51,
+ "learning_rate": 0.00031690705128205126,
+ "loss": 0.6818,
+ "step": 457
+ },
+ {
+ "epoch": 11.54,
+ "learning_rate": 0.00031650641025641026,
+ "loss": 0.692,
+ "step": 458
+ },
+ {
+ "epoch": 11.57,
+ "learning_rate": 0.0003161057692307692,
+ "loss": 0.7028,
+ "step": 459
+ },
+ {
+ "epoch": 11.59,
+ "learning_rate": 0.0003157051282051282,
+ "loss": 0.7009,
+ "step": 460
+ },
+ {
+ "epoch": 11.62,
+ "learning_rate": 0.00031530448717948716,
+ "loss": 0.7325,
+ "step": 461
+ },
+ {
+ "epoch": 11.64,
+ "learning_rate": 0.00031490384615384616,
+ "loss": 0.6738,
+ "step": 462
+ },
+ {
+ "epoch": 11.67,
+ "learning_rate": 0.0003145032051282051,
+ "loss": 0.7008,
+ "step": 463
+ },
+ {
+ "epoch": 11.69,
+ "learning_rate": 0.0003141025641025641,
+ "loss": 0.6838,
+ "step": 464
+ },
+ {
+ "epoch": 11.72,
+ "learning_rate": 0.00031370192307692306,
+ "loss": 0.726,
+ "step": 465
+ },
+ {
+ "epoch": 11.74,
+ "learning_rate": 0.00031330128205128206,
+ "loss": 0.7167,
+ "step": 466
+ },
+ {
+ "epoch": 11.77,
+ "learning_rate": 0.000312900641025641,
+ "loss": 0.7236,
+ "step": 467
+ },
+ {
+ "epoch": 11.79,
+ "learning_rate": 0.0003125,
+ "loss": 0.7076,
+ "step": 468
+ },
+ {
+ "epoch": 11.82,
+ "learning_rate": 0.000312099358974359,
+ "loss": 0.7415,
+ "step": 469
+ },
+ {
+ "epoch": 11.84,
+ "learning_rate": 0.00031169871794871796,
+ "loss": 0.7313,
+ "step": 470
+ },
+ {
+ "epoch": 11.87,
+ "learning_rate": 0.00031129807692307696,
+ "loss": 0.7403,
+ "step": 471
+ },
+ {
+ "epoch": 11.89,
+ "learning_rate": 0.0003108974358974359,
+ "loss": 0.7145,
+ "step": 472
+ },
+ {
+ "epoch": 11.92,
+ "learning_rate": 0.0003104967948717949,
+ "loss": 0.7532,
+ "step": 473
+ },
+ {
+ "epoch": 11.94,
+ "learning_rate": 0.00031009615384615385,
+ "loss": 0.7326,
+ "step": 474
+ },
+ {
+ "epoch": 11.97,
+ "learning_rate": 0.00030969551282051286,
+ "loss": 0.7699,
+ "step": 475
+ },
+ {
+ "epoch": 11.99,
+ "learning_rate": 0.0003092948717948718,
+ "loss": 0.7251,
+ "step": 476
+ },
+ {
+ "epoch": 12.02,
+ "learning_rate": 0.0003088942307692308,
+ "loss": 0.5773,
+ "step": 477
+ },
+ {
+ "epoch": 12.04,
+ "learning_rate": 0.00030849358974358975,
+ "loss": 0.5288,
+ "step": 478
+ },
+ {
+ "epoch": 12.07,
+ "learning_rate": 0.00030809294871794876,
+ "loss": 0.5312,
+ "step": 479
+ },
+ {
+ "epoch": 12.09,
+ "learning_rate": 0.0003076923076923077,
+ "loss": 0.5445,
+ "step": 480
+ },
+ {
+ "epoch": 12.12,
+ "learning_rate": 0.0003072916666666667,
+ "loss": 0.5218,
+ "step": 481
+ },
+ {
+ "epoch": 12.14,
+ "learning_rate": 0.00030689102564102565,
+ "loss": 0.5311,
+ "step": 482
+ },
+ {
+ "epoch": 12.17,
+ "learning_rate": 0.00030649038461538465,
+ "loss": 0.5541,
+ "step": 483
+ },
+ {
+ "epoch": 12.2,
+ "learning_rate": 0.0003060897435897436,
+ "loss": 0.5242,
+ "step": 484
+ },
+ {
+ "epoch": 12.22,
+ "learning_rate": 0.0003056891025641026,
+ "loss": 0.5302,
+ "step": 485
+ },
+ {
+ "epoch": 12.25,
+ "learning_rate": 0.00030528846153846155,
+ "loss": 0.536,
+ "step": 486
+ },
+ {
+ "epoch": 12.27,
+ "learning_rate": 0.00030488782051282055,
+ "loss": 0.5545,
+ "step": 487
+ },
+ {
+ "epoch": 12.3,
+ "learning_rate": 0.00030448717948717945,
+ "loss": 0.5476,
+ "step": 488
+ },
+ {
+ "epoch": 12.32,
+ "learning_rate": 0.00030408653846153845,
+ "loss": 0.5528,
+ "step": 489
+ },
+ {
+ "epoch": 12.35,
+ "learning_rate": 0.0003036858974358974,
+ "loss": 0.5633,
+ "step": 490
+ },
+ {
+ "epoch": 12.37,
+ "learning_rate": 0.0003032852564102564,
+ "loss": 0.571,
+ "step": 491
+ },
+ {
+ "epoch": 12.4,
+ "learning_rate": 0.00030288461538461535,
+ "loss": 0.5637,
+ "step": 492
+ },
+ {
+ "epoch": 12.42,
+ "learning_rate": 0.00030248397435897435,
+ "loss": 0.5613,
+ "step": 493
+ },
+ {
+ "epoch": 12.45,
+ "learning_rate": 0.0003020833333333333,
+ "loss": 0.5999,
+ "step": 494
+ },
+ {
+ "epoch": 12.47,
+ "learning_rate": 0.0003016826923076923,
+ "loss": 0.5633,
+ "step": 495
+ },
+ {
+ "epoch": 12.5,
+ "learning_rate": 0.0003012820512820513,
+ "loss": 0.5811,
+ "step": 496
+ },
+ {
+ "epoch": 12.52,
+ "learning_rate": 0.00030088141025641025,
+ "loss": 0.5711,
+ "step": 497
+ },
+ {
+ "epoch": 12.55,
+ "learning_rate": 0.00030048076923076925,
+ "loss": 0.5669,
+ "step": 498
+ },
+ {
+ "epoch": 12.57,
+ "learning_rate": 0.0003000801282051282,
+ "loss": 0.5424,
+ "step": 499
+ },
+ {
+ "epoch": 12.6,
+ "learning_rate": 0.0002996794871794872,
+ "loss": 0.5592,
+ "step": 500
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 2.8126830046608384e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-500/training_args.bin b/checkpoint-500/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-500/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-600/README.md b/checkpoint-600/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-600/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-600/adapter_config.json b/checkpoint-600/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-600/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-600/adapter_model.bin b/checkpoint-600/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..250d720dbcf3eea43245c3827e46055c9432a65f
--- /dev/null
+++ b/checkpoint-600/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d160518fcf48651459c854c2e492ae94e8830d49081511ef849075147543e998
+size 39409357
diff --git a/checkpoint-600/optimizer.pt b/checkpoint-600/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..ec4738c12656e4779c7cab79352567ac53850880
--- /dev/null
+++ b/checkpoint-600/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c85aaa2f75b62630aaddc294f773deb540aa6125c9a74f237a63e873cf725ab2
+size 78844421
diff --git a/checkpoint-600/rng_state.pth b/checkpoint-600/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ce32862861da12da23bbb01ee9f59c5232cb111d
--- /dev/null
+++ b/checkpoint-600/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:51f3c17c1e7ce128318be2166e96b6dda4d1ff79c8b44e6fb06488bf0bbfcf8d
+size 14575
diff --git a/checkpoint-600/scheduler.pt b/checkpoint-600/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..3daa4177d7f967f11d16cdf7a99ede1f440a15e3
--- /dev/null
+++ b/checkpoint-600/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b3bfddf5edfe0814387844faf92452bba1dbd2638ae27bbb64cc748103b80338
+size 627
diff --git a/checkpoint-600/special_tokens_map.json b/checkpoint-600/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-600/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-600/tokenizer.json b/checkpoint-600/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-600/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-600/tokenizer_config.json b/checkpoint-600/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-600/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-600/trainer_state.json b/checkpoint-600/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..8de40de7d03b377d5100a86b991410bcd1b78d5a
--- /dev/null
+++ b/checkpoint-600/trainer_state.json
@@ -0,0 +1,3619 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 15.118110236220472,
+ "eval_steps": 500,
+ "global_step": 600,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ },
+ {
+ "epoch": 10.1,
+ "learning_rate": 0.00033934294871794873,
+ "loss": 0.7635,
+ "step": 401
+ },
+ {
+ "epoch": 10.13,
+ "learning_rate": 0.00033894230769230773,
+ "loss": 0.7713,
+ "step": 402
+ },
+ {
+ "epoch": 10.15,
+ "learning_rate": 0.0003385416666666667,
+ "loss": 0.827,
+ "step": 403
+ },
+ {
+ "epoch": 10.18,
+ "learning_rate": 0.0003381410256410257,
+ "loss": 0.7844,
+ "step": 404
+ },
+ {
+ "epoch": 10.2,
+ "learning_rate": 0.00033774038461538463,
+ "loss": 0.7839,
+ "step": 405
+ },
+ {
+ "epoch": 10.23,
+ "learning_rate": 0.00033733974358974363,
+ "loss": 0.7702,
+ "step": 406
+ },
+ {
+ "epoch": 10.26,
+ "learning_rate": 0.0003369391025641026,
+ "loss": 0.7855,
+ "step": 407
+ },
+ {
+ "epoch": 10.28,
+ "learning_rate": 0.0003365384615384616,
+ "loss": 0.7669,
+ "step": 408
+ },
+ {
+ "epoch": 10.31,
+ "learning_rate": 0.0003361378205128205,
+ "loss": 0.8054,
+ "step": 409
+ },
+ {
+ "epoch": 10.33,
+ "learning_rate": 0.0003357371794871795,
+ "loss": 0.7792,
+ "step": 410
+ },
+ {
+ "epoch": 10.36,
+ "learning_rate": 0.0003353365384615384,
+ "loss": 0.7986,
+ "step": 411
+ },
+ {
+ "epoch": 10.38,
+ "learning_rate": 0.0003349358974358974,
+ "loss": 0.8269,
+ "step": 412
+ },
+ {
+ "epoch": 10.41,
+ "learning_rate": 0.00033453525641025637,
+ "loss": 0.7923,
+ "step": 413
+ },
+ {
+ "epoch": 10.43,
+ "learning_rate": 0.00033413461538461537,
+ "loss": 0.8242,
+ "step": 414
+ },
+ {
+ "epoch": 10.46,
+ "learning_rate": 0.0003337339743589743,
+ "loss": 0.8496,
+ "step": 415
+ },
+ {
+ "epoch": 10.48,
+ "learning_rate": 0.0003333333333333333,
+ "loss": 0.8236,
+ "step": 416
+ },
+ {
+ "epoch": 10.51,
+ "learning_rate": 0.0003329326923076923,
+ "loss": 0.8685,
+ "step": 417
+ },
+ {
+ "epoch": 10.53,
+ "learning_rate": 0.00033253205128205127,
+ "loss": 0.8542,
+ "step": 418
+ },
+ {
+ "epoch": 10.56,
+ "learning_rate": 0.0003321314102564103,
+ "loss": 0.8185,
+ "step": 419
+ },
+ {
+ "epoch": 10.58,
+ "learning_rate": 0.0003317307692307692,
+ "loss": 0.8477,
+ "step": 420
+ },
+ {
+ "epoch": 10.61,
+ "learning_rate": 0.0003313301282051282,
+ "loss": 0.8372,
+ "step": 421
+ },
+ {
+ "epoch": 10.63,
+ "learning_rate": 0.00033092948717948717,
+ "loss": 0.8731,
+ "step": 422
+ },
+ {
+ "epoch": 10.66,
+ "learning_rate": 0.00033052884615384617,
+ "loss": 0.8322,
+ "step": 423
+ },
+ {
+ "epoch": 10.68,
+ "learning_rate": 0.0003301282051282051,
+ "loss": 0.8514,
+ "step": 424
+ },
+ {
+ "epoch": 10.71,
+ "learning_rate": 0.0003297275641025641,
+ "loss": 0.8564,
+ "step": 425
+ },
+ {
+ "epoch": 10.73,
+ "learning_rate": 0.00032932692307692307,
+ "loss": 0.8278,
+ "step": 426
+ },
+ {
+ "epoch": 10.76,
+ "learning_rate": 0.00032892628205128207,
+ "loss": 0.8408,
+ "step": 427
+ },
+ {
+ "epoch": 10.78,
+ "learning_rate": 0.000328525641025641,
+ "loss": 0.8438,
+ "step": 428
+ },
+ {
+ "epoch": 10.81,
+ "learning_rate": 0.000328125,
+ "loss": 0.8862,
+ "step": 429
+ },
+ {
+ "epoch": 10.83,
+ "learning_rate": 0.00032772435897435897,
+ "loss": 0.8631,
+ "step": 430
+ },
+ {
+ "epoch": 10.86,
+ "learning_rate": 0.00032732371794871797,
+ "loss": 0.883,
+ "step": 431
+ },
+ {
+ "epoch": 10.89,
+ "learning_rate": 0.0003269230769230769,
+ "loss": 0.8644,
+ "step": 432
+ },
+ {
+ "epoch": 10.91,
+ "learning_rate": 0.0003265224358974359,
+ "loss": 0.8907,
+ "step": 433
+ },
+ {
+ "epoch": 10.94,
+ "learning_rate": 0.00032612179487179487,
+ "loss": 0.9141,
+ "step": 434
+ },
+ {
+ "epoch": 10.96,
+ "learning_rate": 0.00032572115384615387,
+ "loss": 0.8877,
+ "step": 435
+ },
+ {
+ "epoch": 10.99,
+ "learning_rate": 0.0003253205128205128,
+ "loss": 0.8521,
+ "step": 436
+ },
+ {
+ "epoch": 11.01,
+ "learning_rate": 0.0003249198717948718,
+ "loss": 0.7842,
+ "step": 437
+ },
+ {
+ "epoch": 11.04,
+ "learning_rate": 0.00032451923076923077,
+ "loss": 0.6516,
+ "step": 438
+ },
+ {
+ "epoch": 11.06,
+ "learning_rate": 0.00032411858974358977,
+ "loss": 0.6549,
+ "step": 439
+ },
+ {
+ "epoch": 11.09,
+ "learning_rate": 0.0003237179487179487,
+ "loss": 0.6736,
+ "step": 440
+ },
+ {
+ "epoch": 11.11,
+ "learning_rate": 0.0003233173076923077,
+ "loss": 0.6625,
+ "step": 441
+ },
+ {
+ "epoch": 11.14,
+ "learning_rate": 0.0003229166666666667,
+ "loss": 0.6466,
+ "step": 442
+ },
+ {
+ "epoch": 11.16,
+ "learning_rate": 0.00032251602564102567,
+ "loss": 0.6362,
+ "step": 443
+ },
+ {
+ "epoch": 11.19,
+ "learning_rate": 0.00032211538461538467,
+ "loss": 0.674,
+ "step": 444
+ },
+ {
+ "epoch": 11.21,
+ "learning_rate": 0.0003217147435897436,
+ "loss": 0.6509,
+ "step": 445
+ },
+ {
+ "epoch": 11.24,
+ "learning_rate": 0.0003213141025641026,
+ "loss": 0.618,
+ "step": 446
+ },
+ {
+ "epoch": 11.26,
+ "learning_rate": 0.00032091346153846156,
+ "loss": 0.6654,
+ "step": 447
+ },
+ {
+ "epoch": 11.29,
+ "learning_rate": 0.00032051282051282057,
+ "loss": 0.6632,
+ "step": 448
+ },
+ {
+ "epoch": 11.31,
+ "learning_rate": 0.00032011217948717946,
+ "loss": 0.6599,
+ "step": 449
+ },
+ {
+ "epoch": 11.34,
+ "learning_rate": 0.00031971153846153846,
+ "loss": 0.6537,
+ "step": 450
+ },
+ {
+ "epoch": 11.36,
+ "learning_rate": 0.0003193108974358974,
+ "loss": 0.678,
+ "step": 451
+ },
+ {
+ "epoch": 11.39,
+ "learning_rate": 0.0003189102564102564,
+ "loss": 0.684,
+ "step": 452
+ },
+ {
+ "epoch": 11.41,
+ "learning_rate": 0.00031850961538461536,
+ "loss": 0.6419,
+ "step": 453
+ },
+ {
+ "epoch": 11.44,
+ "learning_rate": 0.00031810897435897436,
+ "loss": 0.6614,
+ "step": 454
+ },
+ {
+ "epoch": 11.46,
+ "learning_rate": 0.0003177083333333333,
+ "loss": 0.7037,
+ "step": 455
+ },
+ {
+ "epoch": 11.49,
+ "learning_rate": 0.0003173076923076923,
+ "loss": 0.6944,
+ "step": 456
+ },
+ {
+ "epoch": 11.51,
+ "learning_rate": 0.00031690705128205126,
+ "loss": 0.6818,
+ "step": 457
+ },
+ {
+ "epoch": 11.54,
+ "learning_rate": 0.00031650641025641026,
+ "loss": 0.692,
+ "step": 458
+ },
+ {
+ "epoch": 11.57,
+ "learning_rate": 0.0003161057692307692,
+ "loss": 0.7028,
+ "step": 459
+ },
+ {
+ "epoch": 11.59,
+ "learning_rate": 0.0003157051282051282,
+ "loss": 0.7009,
+ "step": 460
+ },
+ {
+ "epoch": 11.62,
+ "learning_rate": 0.00031530448717948716,
+ "loss": 0.7325,
+ "step": 461
+ },
+ {
+ "epoch": 11.64,
+ "learning_rate": 0.00031490384615384616,
+ "loss": 0.6738,
+ "step": 462
+ },
+ {
+ "epoch": 11.67,
+ "learning_rate": 0.0003145032051282051,
+ "loss": 0.7008,
+ "step": 463
+ },
+ {
+ "epoch": 11.69,
+ "learning_rate": 0.0003141025641025641,
+ "loss": 0.6838,
+ "step": 464
+ },
+ {
+ "epoch": 11.72,
+ "learning_rate": 0.00031370192307692306,
+ "loss": 0.726,
+ "step": 465
+ },
+ {
+ "epoch": 11.74,
+ "learning_rate": 0.00031330128205128206,
+ "loss": 0.7167,
+ "step": 466
+ },
+ {
+ "epoch": 11.77,
+ "learning_rate": 0.000312900641025641,
+ "loss": 0.7236,
+ "step": 467
+ },
+ {
+ "epoch": 11.79,
+ "learning_rate": 0.0003125,
+ "loss": 0.7076,
+ "step": 468
+ },
+ {
+ "epoch": 11.82,
+ "learning_rate": 0.000312099358974359,
+ "loss": 0.7415,
+ "step": 469
+ },
+ {
+ "epoch": 11.84,
+ "learning_rate": 0.00031169871794871796,
+ "loss": 0.7313,
+ "step": 470
+ },
+ {
+ "epoch": 11.87,
+ "learning_rate": 0.00031129807692307696,
+ "loss": 0.7403,
+ "step": 471
+ },
+ {
+ "epoch": 11.89,
+ "learning_rate": 0.0003108974358974359,
+ "loss": 0.7145,
+ "step": 472
+ },
+ {
+ "epoch": 11.92,
+ "learning_rate": 0.0003104967948717949,
+ "loss": 0.7532,
+ "step": 473
+ },
+ {
+ "epoch": 11.94,
+ "learning_rate": 0.00031009615384615385,
+ "loss": 0.7326,
+ "step": 474
+ },
+ {
+ "epoch": 11.97,
+ "learning_rate": 0.00030969551282051286,
+ "loss": 0.7699,
+ "step": 475
+ },
+ {
+ "epoch": 11.99,
+ "learning_rate": 0.0003092948717948718,
+ "loss": 0.7251,
+ "step": 476
+ },
+ {
+ "epoch": 12.02,
+ "learning_rate": 0.0003088942307692308,
+ "loss": 0.5773,
+ "step": 477
+ },
+ {
+ "epoch": 12.04,
+ "learning_rate": 0.00030849358974358975,
+ "loss": 0.5288,
+ "step": 478
+ },
+ {
+ "epoch": 12.07,
+ "learning_rate": 0.00030809294871794876,
+ "loss": 0.5312,
+ "step": 479
+ },
+ {
+ "epoch": 12.09,
+ "learning_rate": 0.0003076923076923077,
+ "loss": 0.5445,
+ "step": 480
+ },
+ {
+ "epoch": 12.12,
+ "learning_rate": 0.0003072916666666667,
+ "loss": 0.5218,
+ "step": 481
+ },
+ {
+ "epoch": 12.14,
+ "learning_rate": 0.00030689102564102565,
+ "loss": 0.5311,
+ "step": 482
+ },
+ {
+ "epoch": 12.17,
+ "learning_rate": 0.00030649038461538465,
+ "loss": 0.5541,
+ "step": 483
+ },
+ {
+ "epoch": 12.2,
+ "learning_rate": 0.0003060897435897436,
+ "loss": 0.5242,
+ "step": 484
+ },
+ {
+ "epoch": 12.22,
+ "learning_rate": 0.0003056891025641026,
+ "loss": 0.5302,
+ "step": 485
+ },
+ {
+ "epoch": 12.25,
+ "learning_rate": 0.00030528846153846155,
+ "loss": 0.536,
+ "step": 486
+ },
+ {
+ "epoch": 12.27,
+ "learning_rate": 0.00030488782051282055,
+ "loss": 0.5545,
+ "step": 487
+ },
+ {
+ "epoch": 12.3,
+ "learning_rate": 0.00030448717948717945,
+ "loss": 0.5476,
+ "step": 488
+ },
+ {
+ "epoch": 12.32,
+ "learning_rate": 0.00030408653846153845,
+ "loss": 0.5528,
+ "step": 489
+ },
+ {
+ "epoch": 12.35,
+ "learning_rate": 0.0003036858974358974,
+ "loss": 0.5633,
+ "step": 490
+ },
+ {
+ "epoch": 12.37,
+ "learning_rate": 0.0003032852564102564,
+ "loss": 0.571,
+ "step": 491
+ },
+ {
+ "epoch": 12.4,
+ "learning_rate": 0.00030288461538461535,
+ "loss": 0.5637,
+ "step": 492
+ },
+ {
+ "epoch": 12.42,
+ "learning_rate": 0.00030248397435897435,
+ "loss": 0.5613,
+ "step": 493
+ },
+ {
+ "epoch": 12.45,
+ "learning_rate": 0.0003020833333333333,
+ "loss": 0.5999,
+ "step": 494
+ },
+ {
+ "epoch": 12.47,
+ "learning_rate": 0.0003016826923076923,
+ "loss": 0.5633,
+ "step": 495
+ },
+ {
+ "epoch": 12.5,
+ "learning_rate": 0.0003012820512820513,
+ "loss": 0.5811,
+ "step": 496
+ },
+ {
+ "epoch": 12.52,
+ "learning_rate": 0.00030088141025641025,
+ "loss": 0.5711,
+ "step": 497
+ },
+ {
+ "epoch": 12.55,
+ "learning_rate": 0.00030048076923076925,
+ "loss": 0.5669,
+ "step": 498
+ },
+ {
+ "epoch": 12.57,
+ "learning_rate": 0.0003000801282051282,
+ "loss": 0.5424,
+ "step": 499
+ },
+ {
+ "epoch": 12.6,
+ "learning_rate": 0.0002996794871794872,
+ "loss": 0.5592,
+ "step": 500
+ },
+ {
+ "epoch": 12.62,
+ "learning_rate": 0.00029927884615384614,
+ "loss": 0.5645,
+ "step": 501
+ },
+ {
+ "epoch": 12.65,
+ "learning_rate": 0.00029887820512820515,
+ "loss": 0.6089,
+ "step": 502
+ },
+ {
+ "epoch": 12.67,
+ "learning_rate": 0.0002984775641025641,
+ "loss": 0.57,
+ "step": 503
+ },
+ {
+ "epoch": 12.7,
+ "learning_rate": 0.0002980769230769231,
+ "loss": 0.57,
+ "step": 504
+ },
+ {
+ "epoch": 12.72,
+ "learning_rate": 0.00029767628205128204,
+ "loss": 0.5511,
+ "step": 505
+ },
+ {
+ "epoch": 12.75,
+ "learning_rate": 0.00029727564102564105,
+ "loss": 0.5765,
+ "step": 506
+ },
+ {
+ "epoch": 12.77,
+ "learning_rate": 0.000296875,
+ "loss": 0.5896,
+ "step": 507
+ },
+ {
+ "epoch": 12.8,
+ "learning_rate": 0.000296474358974359,
+ "loss": 0.6329,
+ "step": 508
+ },
+ {
+ "epoch": 12.83,
+ "learning_rate": 0.00029607371794871794,
+ "loss": 0.5682,
+ "step": 509
+ },
+ {
+ "epoch": 12.85,
+ "learning_rate": 0.00029567307692307694,
+ "loss": 0.58,
+ "step": 510
+ },
+ {
+ "epoch": 12.88,
+ "learning_rate": 0.0002952724358974359,
+ "loss": 0.6014,
+ "step": 511
+ },
+ {
+ "epoch": 12.9,
+ "learning_rate": 0.0002948717948717949,
+ "loss": 0.6082,
+ "step": 512
+ },
+ {
+ "epoch": 12.93,
+ "learning_rate": 0.00029447115384615384,
+ "loss": 0.6392,
+ "step": 513
+ },
+ {
+ "epoch": 12.95,
+ "learning_rate": 0.00029407051282051284,
+ "loss": 0.609,
+ "step": 514
+ },
+ {
+ "epoch": 12.98,
+ "learning_rate": 0.0002936698717948718,
+ "loss": 0.6247,
+ "step": 515
+ },
+ {
+ "epoch": 13.0,
+ "learning_rate": 0.0002932692307692308,
+ "loss": 0.5943,
+ "step": 516
+ },
+ {
+ "epoch": 13.03,
+ "learning_rate": 0.00029286858974358974,
+ "loss": 0.4499,
+ "step": 517
+ },
+ {
+ "epoch": 13.05,
+ "learning_rate": 0.00029246794871794874,
+ "loss": 0.4412,
+ "step": 518
+ },
+ {
+ "epoch": 13.08,
+ "learning_rate": 0.0002920673076923077,
+ "loss": 0.4409,
+ "step": 519
+ },
+ {
+ "epoch": 13.1,
+ "learning_rate": 0.0002916666666666667,
+ "loss": 0.4505,
+ "step": 520
+ },
+ {
+ "epoch": 13.13,
+ "learning_rate": 0.0002912660256410257,
+ "loss": 0.4489,
+ "step": 521
+ },
+ {
+ "epoch": 13.15,
+ "learning_rate": 0.00029086538461538464,
+ "loss": 0.4235,
+ "step": 522
+ },
+ {
+ "epoch": 13.18,
+ "learning_rate": 0.00029046474358974364,
+ "loss": 0.4276,
+ "step": 523
+ },
+ {
+ "epoch": 13.2,
+ "learning_rate": 0.0002900641025641026,
+ "loss": 0.4544,
+ "step": 524
+ },
+ {
+ "epoch": 13.23,
+ "learning_rate": 0.0002896634615384616,
+ "loss": 0.4625,
+ "step": 525
+ },
+ {
+ "epoch": 13.25,
+ "learning_rate": 0.00028926282051282054,
+ "loss": 0.4266,
+ "step": 526
+ },
+ {
+ "epoch": 13.28,
+ "learning_rate": 0.0002888621794871795,
+ "loss": 0.4459,
+ "step": 527
+ },
+ {
+ "epoch": 13.3,
+ "learning_rate": 0.00028846153846153843,
+ "loss": 0.4261,
+ "step": 528
+ },
+ {
+ "epoch": 13.33,
+ "learning_rate": 0.00028806089743589744,
+ "loss": 0.4376,
+ "step": 529
+ },
+ {
+ "epoch": 13.35,
+ "learning_rate": 0.0002876602564102564,
+ "loss": 0.4418,
+ "step": 530
+ },
+ {
+ "epoch": 13.38,
+ "learning_rate": 0.0002872596153846154,
+ "loss": 0.4667,
+ "step": 531
+ },
+ {
+ "epoch": 13.4,
+ "learning_rate": 0.00028685897435897433,
+ "loss": 0.4504,
+ "step": 532
+ },
+ {
+ "epoch": 13.43,
+ "learning_rate": 0.00028645833333333333,
+ "loss": 0.4691,
+ "step": 533
+ },
+ {
+ "epoch": 13.46,
+ "learning_rate": 0.0002860576923076923,
+ "loss": 0.4684,
+ "step": 534
+ },
+ {
+ "epoch": 13.48,
+ "learning_rate": 0.0002856570512820513,
+ "loss": 0.467,
+ "step": 535
+ },
+ {
+ "epoch": 13.51,
+ "learning_rate": 0.00028525641025641023,
+ "loss": 0.4716,
+ "step": 536
+ },
+ {
+ "epoch": 13.53,
+ "learning_rate": 0.00028485576923076923,
+ "loss": 0.4619,
+ "step": 537
+ },
+ {
+ "epoch": 13.56,
+ "learning_rate": 0.0002844551282051282,
+ "loss": 0.4775,
+ "step": 538
+ },
+ {
+ "epoch": 13.58,
+ "learning_rate": 0.0002840544871794872,
+ "loss": 0.4568,
+ "step": 539
+ },
+ {
+ "epoch": 13.61,
+ "learning_rate": 0.00028365384615384613,
+ "loss": 0.4763,
+ "step": 540
+ },
+ {
+ "epoch": 13.63,
+ "learning_rate": 0.00028325320512820513,
+ "loss": 0.466,
+ "step": 541
+ },
+ {
+ "epoch": 13.66,
+ "learning_rate": 0.0002828525641025641,
+ "loss": 0.4543,
+ "step": 542
+ },
+ {
+ "epoch": 13.68,
+ "learning_rate": 0.0002824519230769231,
+ "loss": 0.4666,
+ "step": 543
+ },
+ {
+ "epoch": 13.71,
+ "learning_rate": 0.00028205128205128203,
+ "loss": 0.489,
+ "step": 544
+ },
+ {
+ "epoch": 13.73,
+ "learning_rate": 0.00028165064102564103,
+ "loss": 0.491,
+ "step": 545
+ },
+ {
+ "epoch": 13.76,
+ "learning_rate": 0.00028125000000000003,
+ "loss": 0.4847,
+ "step": 546
+ },
+ {
+ "epoch": 13.78,
+ "learning_rate": 0.000280849358974359,
+ "loss": 0.4899,
+ "step": 547
+ },
+ {
+ "epoch": 13.81,
+ "learning_rate": 0.000280448717948718,
+ "loss": 0.5169,
+ "step": 548
+ },
+ {
+ "epoch": 13.83,
+ "learning_rate": 0.00028004807692307693,
+ "loss": 0.4968,
+ "step": 549
+ },
+ {
+ "epoch": 13.86,
+ "learning_rate": 0.00027964743589743593,
+ "loss": 0.4774,
+ "step": 550
+ },
+ {
+ "epoch": 13.88,
+ "learning_rate": 0.0002792467948717949,
+ "loss": 0.5105,
+ "step": 551
+ },
+ {
+ "epoch": 13.91,
+ "learning_rate": 0.0002788461538461539,
+ "loss": 0.4997,
+ "step": 552
+ },
+ {
+ "epoch": 13.93,
+ "learning_rate": 0.00027844551282051283,
+ "loss": 0.5088,
+ "step": 553
+ },
+ {
+ "epoch": 13.96,
+ "learning_rate": 0.00027804487179487183,
+ "loss": 0.5091,
+ "step": 554
+ },
+ {
+ "epoch": 13.98,
+ "learning_rate": 0.0002776442307692308,
+ "loss": 0.532,
+ "step": 555
+ },
+ {
+ "epoch": 14.01,
+ "learning_rate": 0.0002772435897435898,
+ "loss": 0.4487,
+ "step": 556
+ },
+ {
+ "epoch": 14.03,
+ "learning_rate": 0.00027684294871794873,
+ "loss": 0.3525,
+ "step": 557
+ },
+ {
+ "epoch": 14.06,
+ "learning_rate": 0.00027644230769230773,
+ "loss": 0.3715,
+ "step": 558
+ },
+ {
+ "epoch": 14.09,
+ "learning_rate": 0.0002760416666666667,
+ "loss": 0.3547,
+ "step": 559
+ },
+ {
+ "epoch": 14.11,
+ "learning_rate": 0.0002756410256410257,
+ "loss": 0.3862,
+ "step": 560
+ },
+ {
+ "epoch": 14.14,
+ "learning_rate": 0.0002752403846153846,
+ "loss": 0.3514,
+ "step": 561
+ },
+ {
+ "epoch": 14.16,
+ "learning_rate": 0.00027483974358974363,
+ "loss": 0.362,
+ "step": 562
+ },
+ {
+ "epoch": 14.19,
+ "learning_rate": 0.0002744391025641026,
+ "loss": 0.3457,
+ "step": 563
+ },
+ {
+ "epoch": 14.21,
+ "learning_rate": 0.0002740384615384616,
+ "loss": 0.3662,
+ "step": 564
+ },
+ {
+ "epoch": 14.24,
+ "learning_rate": 0.0002736378205128205,
+ "loss": 0.3746,
+ "step": 565
+ },
+ {
+ "epoch": 14.26,
+ "learning_rate": 0.00027323717948717947,
+ "loss": 0.3388,
+ "step": 566
+ },
+ {
+ "epoch": 14.29,
+ "learning_rate": 0.0002728365384615384,
+ "loss": 0.3824,
+ "step": 567
+ },
+ {
+ "epoch": 14.31,
+ "learning_rate": 0.0002724358974358974,
+ "loss": 0.3584,
+ "step": 568
+ },
+ {
+ "epoch": 14.34,
+ "learning_rate": 0.00027203525641025637,
+ "loss": 0.3544,
+ "step": 569
+ },
+ {
+ "epoch": 14.36,
+ "learning_rate": 0.00027163461538461537,
+ "loss": 0.3721,
+ "step": 570
+ },
+ {
+ "epoch": 14.39,
+ "learning_rate": 0.0002712339743589743,
+ "loss": 0.3769,
+ "step": 571
+ },
+ {
+ "epoch": 14.41,
+ "learning_rate": 0.0002708333333333333,
+ "loss": 0.3872,
+ "step": 572
+ },
+ {
+ "epoch": 14.44,
+ "learning_rate": 0.0002704326923076923,
+ "loss": 0.363,
+ "step": 573
+ },
+ {
+ "epoch": 14.46,
+ "learning_rate": 0.00027003205128205127,
+ "loss": 0.3842,
+ "step": 574
+ },
+ {
+ "epoch": 14.49,
+ "learning_rate": 0.00026963141025641027,
+ "loss": 0.3737,
+ "step": 575
+ },
+ {
+ "epoch": 14.51,
+ "learning_rate": 0.0002692307692307692,
+ "loss": 0.3981,
+ "step": 576
+ },
+ {
+ "epoch": 14.54,
+ "learning_rate": 0.0002688301282051282,
+ "loss": 0.3656,
+ "step": 577
+ },
+ {
+ "epoch": 14.56,
+ "learning_rate": 0.00026842948717948717,
+ "loss": 0.3852,
+ "step": 578
+ },
+ {
+ "epoch": 14.59,
+ "learning_rate": 0.00026802884615384617,
+ "loss": 0.3865,
+ "step": 579
+ },
+ {
+ "epoch": 14.61,
+ "learning_rate": 0.0002676282051282051,
+ "loss": 0.3747,
+ "step": 580
+ },
+ {
+ "epoch": 14.64,
+ "learning_rate": 0.0002672275641025641,
+ "loss": 0.3729,
+ "step": 581
+ },
+ {
+ "epoch": 14.66,
+ "learning_rate": 0.00026682692307692307,
+ "loss": 0.3785,
+ "step": 582
+ },
+ {
+ "epoch": 14.69,
+ "learning_rate": 0.00026642628205128207,
+ "loss": 0.4018,
+ "step": 583
+ },
+ {
+ "epoch": 14.71,
+ "learning_rate": 0.000266025641025641,
+ "loss": 0.3854,
+ "step": 584
+ },
+ {
+ "epoch": 14.74,
+ "learning_rate": 0.000265625,
+ "loss": 0.4161,
+ "step": 585
+ },
+ {
+ "epoch": 14.77,
+ "learning_rate": 0.00026522435897435897,
+ "loss": 0.4093,
+ "step": 586
+ },
+ {
+ "epoch": 14.79,
+ "learning_rate": 0.00026482371794871797,
+ "loss": 0.4126,
+ "step": 587
+ },
+ {
+ "epoch": 14.82,
+ "learning_rate": 0.0002644230769230769,
+ "loss": 0.3977,
+ "step": 588
+ },
+ {
+ "epoch": 14.84,
+ "learning_rate": 0.0002640224358974359,
+ "loss": 0.4286,
+ "step": 589
+ },
+ {
+ "epoch": 14.87,
+ "learning_rate": 0.00026362179487179487,
+ "loss": 0.4371,
+ "step": 590
+ },
+ {
+ "epoch": 14.89,
+ "learning_rate": 0.00026322115384615387,
+ "loss": 0.4066,
+ "step": 591
+ },
+ {
+ "epoch": 14.92,
+ "learning_rate": 0.0002628205128205128,
+ "loss": 0.433,
+ "step": 592
+ },
+ {
+ "epoch": 14.94,
+ "learning_rate": 0.0002624198717948718,
+ "loss": 0.4361,
+ "step": 593
+ },
+ {
+ "epoch": 14.97,
+ "learning_rate": 0.00026201923076923076,
+ "loss": 0.4276,
+ "step": 594
+ },
+ {
+ "epoch": 14.99,
+ "learning_rate": 0.00026161858974358977,
+ "loss": 0.4401,
+ "step": 595
+ },
+ {
+ "epoch": 15.02,
+ "learning_rate": 0.0002612179487179487,
+ "loss": 0.3514,
+ "step": 596
+ },
+ {
+ "epoch": 15.04,
+ "learning_rate": 0.0002608173076923077,
+ "loss": 0.306,
+ "step": 597
+ },
+ {
+ "epoch": 15.07,
+ "learning_rate": 0.0002604166666666667,
+ "loss": 0.2862,
+ "step": 598
+ },
+ {
+ "epoch": 15.09,
+ "learning_rate": 0.00026001602564102566,
+ "loss": 0.3042,
+ "step": 599
+ },
+ {
+ "epoch": 15.12,
+ "learning_rate": 0.00025961538461538467,
+ "loss": 0.2978,
+ "step": 600
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 3.374598554368205e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-600/training_args.bin b/checkpoint-600/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-600/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-700/README.md b/checkpoint-700/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-700/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-700/adapter_config.json b/checkpoint-700/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-700/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-700/adapter_model.bin b/checkpoint-700/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..719c7447b765f985fe07a808ae37318550eb44d5
--- /dev/null
+++ b/checkpoint-700/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:10d0ddb7b0ad9e7b24305c9fe81b630c7e311dca3f539c0d1662dd7118c771db
+size 39409357
diff --git a/checkpoint-700/optimizer.pt b/checkpoint-700/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..06c2db93af252f118c760a872cd3dcdfdef01d6c
--- /dev/null
+++ b/checkpoint-700/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:47406f9dc7b92dad5dfc0a371fe784d1ed6cc7b426c6f1a3c00e4e915aa7aff6
+size 78844421
diff --git a/checkpoint-700/rng_state.pth b/checkpoint-700/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f3561d2e8364fc4a00c0d78eca485fdee4dcf81a
--- /dev/null
+++ b/checkpoint-700/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9be2d54ebe54c824e6446e6978ff3aefa7199568ed49c03eaafa5cd13a62e8bf
+size 14575
diff --git a/checkpoint-700/scheduler.pt b/checkpoint-700/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..69993829fa46560d57236dd3310b3fa068023203
--- /dev/null
+++ b/checkpoint-700/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:270b582fa7600883da68f58a70ba401ded3d803d695d8ea5c5db11fbc934334a
+size 627
diff --git a/checkpoint-700/special_tokens_map.json b/checkpoint-700/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-700/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-700/tokenizer.json b/checkpoint-700/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-700/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-700/tokenizer_config.json b/checkpoint-700/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-700/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-700/trainer_state.json b/checkpoint-700/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..1e26931e61031416a078b43e6cec58006957d417
--- /dev/null
+++ b/checkpoint-700/trainer_state.json
@@ -0,0 +1,4219 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 17.637795275590552,
+ "eval_steps": 500,
+ "global_step": 700,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ },
+ {
+ "epoch": 10.1,
+ "learning_rate": 0.00033934294871794873,
+ "loss": 0.7635,
+ "step": 401
+ },
+ {
+ "epoch": 10.13,
+ "learning_rate": 0.00033894230769230773,
+ "loss": 0.7713,
+ "step": 402
+ },
+ {
+ "epoch": 10.15,
+ "learning_rate": 0.0003385416666666667,
+ "loss": 0.827,
+ "step": 403
+ },
+ {
+ "epoch": 10.18,
+ "learning_rate": 0.0003381410256410257,
+ "loss": 0.7844,
+ "step": 404
+ },
+ {
+ "epoch": 10.2,
+ "learning_rate": 0.00033774038461538463,
+ "loss": 0.7839,
+ "step": 405
+ },
+ {
+ "epoch": 10.23,
+ "learning_rate": 0.00033733974358974363,
+ "loss": 0.7702,
+ "step": 406
+ },
+ {
+ "epoch": 10.26,
+ "learning_rate": 0.0003369391025641026,
+ "loss": 0.7855,
+ "step": 407
+ },
+ {
+ "epoch": 10.28,
+ "learning_rate": 0.0003365384615384616,
+ "loss": 0.7669,
+ "step": 408
+ },
+ {
+ "epoch": 10.31,
+ "learning_rate": 0.0003361378205128205,
+ "loss": 0.8054,
+ "step": 409
+ },
+ {
+ "epoch": 10.33,
+ "learning_rate": 0.0003357371794871795,
+ "loss": 0.7792,
+ "step": 410
+ },
+ {
+ "epoch": 10.36,
+ "learning_rate": 0.0003353365384615384,
+ "loss": 0.7986,
+ "step": 411
+ },
+ {
+ "epoch": 10.38,
+ "learning_rate": 0.0003349358974358974,
+ "loss": 0.8269,
+ "step": 412
+ },
+ {
+ "epoch": 10.41,
+ "learning_rate": 0.00033453525641025637,
+ "loss": 0.7923,
+ "step": 413
+ },
+ {
+ "epoch": 10.43,
+ "learning_rate": 0.00033413461538461537,
+ "loss": 0.8242,
+ "step": 414
+ },
+ {
+ "epoch": 10.46,
+ "learning_rate": 0.0003337339743589743,
+ "loss": 0.8496,
+ "step": 415
+ },
+ {
+ "epoch": 10.48,
+ "learning_rate": 0.0003333333333333333,
+ "loss": 0.8236,
+ "step": 416
+ },
+ {
+ "epoch": 10.51,
+ "learning_rate": 0.0003329326923076923,
+ "loss": 0.8685,
+ "step": 417
+ },
+ {
+ "epoch": 10.53,
+ "learning_rate": 0.00033253205128205127,
+ "loss": 0.8542,
+ "step": 418
+ },
+ {
+ "epoch": 10.56,
+ "learning_rate": 0.0003321314102564103,
+ "loss": 0.8185,
+ "step": 419
+ },
+ {
+ "epoch": 10.58,
+ "learning_rate": 0.0003317307692307692,
+ "loss": 0.8477,
+ "step": 420
+ },
+ {
+ "epoch": 10.61,
+ "learning_rate": 0.0003313301282051282,
+ "loss": 0.8372,
+ "step": 421
+ },
+ {
+ "epoch": 10.63,
+ "learning_rate": 0.00033092948717948717,
+ "loss": 0.8731,
+ "step": 422
+ },
+ {
+ "epoch": 10.66,
+ "learning_rate": 0.00033052884615384617,
+ "loss": 0.8322,
+ "step": 423
+ },
+ {
+ "epoch": 10.68,
+ "learning_rate": 0.0003301282051282051,
+ "loss": 0.8514,
+ "step": 424
+ },
+ {
+ "epoch": 10.71,
+ "learning_rate": 0.0003297275641025641,
+ "loss": 0.8564,
+ "step": 425
+ },
+ {
+ "epoch": 10.73,
+ "learning_rate": 0.00032932692307692307,
+ "loss": 0.8278,
+ "step": 426
+ },
+ {
+ "epoch": 10.76,
+ "learning_rate": 0.00032892628205128207,
+ "loss": 0.8408,
+ "step": 427
+ },
+ {
+ "epoch": 10.78,
+ "learning_rate": 0.000328525641025641,
+ "loss": 0.8438,
+ "step": 428
+ },
+ {
+ "epoch": 10.81,
+ "learning_rate": 0.000328125,
+ "loss": 0.8862,
+ "step": 429
+ },
+ {
+ "epoch": 10.83,
+ "learning_rate": 0.00032772435897435897,
+ "loss": 0.8631,
+ "step": 430
+ },
+ {
+ "epoch": 10.86,
+ "learning_rate": 0.00032732371794871797,
+ "loss": 0.883,
+ "step": 431
+ },
+ {
+ "epoch": 10.89,
+ "learning_rate": 0.0003269230769230769,
+ "loss": 0.8644,
+ "step": 432
+ },
+ {
+ "epoch": 10.91,
+ "learning_rate": 0.0003265224358974359,
+ "loss": 0.8907,
+ "step": 433
+ },
+ {
+ "epoch": 10.94,
+ "learning_rate": 0.00032612179487179487,
+ "loss": 0.9141,
+ "step": 434
+ },
+ {
+ "epoch": 10.96,
+ "learning_rate": 0.00032572115384615387,
+ "loss": 0.8877,
+ "step": 435
+ },
+ {
+ "epoch": 10.99,
+ "learning_rate": 0.0003253205128205128,
+ "loss": 0.8521,
+ "step": 436
+ },
+ {
+ "epoch": 11.01,
+ "learning_rate": 0.0003249198717948718,
+ "loss": 0.7842,
+ "step": 437
+ },
+ {
+ "epoch": 11.04,
+ "learning_rate": 0.00032451923076923077,
+ "loss": 0.6516,
+ "step": 438
+ },
+ {
+ "epoch": 11.06,
+ "learning_rate": 0.00032411858974358977,
+ "loss": 0.6549,
+ "step": 439
+ },
+ {
+ "epoch": 11.09,
+ "learning_rate": 0.0003237179487179487,
+ "loss": 0.6736,
+ "step": 440
+ },
+ {
+ "epoch": 11.11,
+ "learning_rate": 0.0003233173076923077,
+ "loss": 0.6625,
+ "step": 441
+ },
+ {
+ "epoch": 11.14,
+ "learning_rate": 0.0003229166666666667,
+ "loss": 0.6466,
+ "step": 442
+ },
+ {
+ "epoch": 11.16,
+ "learning_rate": 0.00032251602564102567,
+ "loss": 0.6362,
+ "step": 443
+ },
+ {
+ "epoch": 11.19,
+ "learning_rate": 0.00032211538461538467,
+ "loss": 0.674,
+ "step": 444
+ },
+ {
+ "epoch": 11.21,
+ "learning_rate": 0.0003217147435897436,
+ "loss": 0.6509,
+ "step": 445
+ },
+ {
+ "epoch": 11.24,
+ "learning_rate": 0.0003213141025641026,
+ "loss": 0.618,
+ "step": 446
+ },
+ {
+ "epoch": 11.26,
+ "learning_rate": 0.00032091346153846156,
+ "loss": 0.6654,
+ "step": 447
+ },
+ {
+ "epoch": 11.29,
+ "learning_rate": 0.00032051282051282057,
+ "loss": 0.6632,
+ "step": 448
+ },
+ {
+ "epoch": 11.31,
+ "learning_rate": 0.00032011217948717946,
+ "loss": 0.6599,
+ "step": 449
+ },
+ {
+ "epoch": 11.34,
+ "learning_rate": 0.00031971153846153846,
+ "loss": 0.6537,
+ "step": 450
+ },
+ {
+ "epoch": 11.36,
+ "learning_rate": 0.0003193108974358974,
+ "loss": 0.678,
+ "step": 451
+ },
+ {
+ "epoch": 11.39,
+ "learning_rate": 0.0003189102564102564,
+ "loss": 0.684,
+ "step": 452
+ },
+ {
+ "epoch": 11.41,
+ "learning_rate": 0.00031850961538461536,
+ "loss": 0.6419,
+ "step": 453
+ },
+ {
+ "epoch": 11.44,
+ "learning_rate": 0.00031810897435897436,
+ "loss": 0.6614,
+ "step": 454
+ },
+ {
+ "epoch": 11.46,
+ "learning_rate": 0.0003177083333333333,
+ "loss": 0.7037,
+ "step": 455
+ },
+ {
+ "epoch": 11.49,
+ "learning_rate": 0.0003173076923076923,
+ "loss": 0.6944,
+ "step": 456
+ },
+ {
+ "epoch": 11.51,
+ "learning_rate": 0.00031690705128205126,
+ "loss": 0.6818,
+ "step": 457
+ },
+ {
+ "epoch": 11.54,
+ "learning_rate": 0.00031650641025641026,
+ "loss": 0.692,
+ "step": 458
+ },
+ {
+ "epoch": 11.57,
+ "learning_rate": 0.0003161057692307692,
+ "loss": 0.7028,
+ "step": 459
+ },
+ {
+ "epoch": 11.59,
+ "learning_rate": 0.0003157051282051282,
+ "loss": 0.7009,
+ "step": 460
+ },
+ {
+ "epoch": 11.62,
+ "learning_rate": 0.00031530448717948716,
+ "loss": 0.7325,
+ "step": 461
+ },
+ {
+ "epoch": 11.64,
+ "learning_rate": 0.00031490384615384616,
+ "loss": 0.6738,
+ "step": 462
+ },
+ {
+ "epoch": 11.67,
+ "learning_rate": 0.0003145032051282051,
+ "loss": 0.7008,
+ "step": 463
+ },
+ {
+ "epoch": 11.69,
+ "learning_rate": 0.0003141025641025641,
+ "loss": 0.6838,
+ "step": 464
+ },
+ {
+ "epoch": 11.72,
+ "learning_rate": 0.00031370192307692306,
+ "loss": 0.726,
+ "step": 465
+ },
+ {
+ "epoch": 11.74,
+ "learning_rate": 0.00031330128205128206,
+ "loss": 0.7167,
+ "step": 466
+ },
+ {
+ "epoch": 11.77,
+ "learning_rate": 0.000312900641025641,
+ "loss": 0.7236,
+ "step": 467
+ },
+ {
+ "epoch": 11.79,
+ "learning_rate": 0.0003125,
+ "loss": 0.7076,
+ "step": 468
+ },
+ {
+ "epoch": 11.82,
+ "learning_rate": 0.000312099358974359,
+ "loss": 0.7415,
+ "step": 469
+ },
+ {
+ "epoch": 11.84,
+ "learning_rate": 0.00031169871794871796,
+ "loss": 0.7313,
+ "step": 470
+ },
+ {
+ "epoch": 11.87,
+ "learning_rate": 0.00031129807692307696,
+ "loss": 0.7403,
+ "step": 471
+ },
+ {
+ "epoch": 11.89,
+ "learning_rate": 0.0003108974358974359,
+ "loss": 0.7145,
+ "step": 472
+ },
+ {
+ "epoch": 11.92,
+ "learning_rate": 0.0003104967948717949,
+ "loss": 0.7532,
+ "step": 473
+ },
+ {
+ "epoch": 11.94,
+ "learning_rate": 0.00031009615384615385,
+ "loss": 0.7326,
+ "step": 474
+ },
+ {
+ "epoch": 11.97,
+ "learning_rate": 0.00030969551282051286,
+ "loss": 0.7699,
+ "step": 475
+ },
+ {
+ "epoch": 11.99,
+ "learning_rate": 0.0003092948717948718,
+ "loss": 0.7251,
+ "step": 476
+ },
+ {
+ "epoch": 12.02,
+ "learning_rate": 0.0003088942307692308,
+ "loss": 0.5773,
+ "step": 477
+ },
+ {
+ "epoch": 12.04,
+ "learning_rate": 0.00030849358974358975,
+ "loss": 0.5288,
+ "step": 478
+ },
+ {
+ "epoch": 12.07,
+ "learning_rate": 0.00030809294871794876,
+ "loss": 0.5312,
+ "step": 479
+ },
+ {
+ "epoch": 12.09,
+ "learning_rate": 0.0003076923076923077,
+ "loss": 0.5445,
+ "step": 480
+ },
+ {
+ "epoch": 12.12,
+ "learning_rate": 0.0003072916666666667,
+ "loss": 0.5218,
+ "step": 481
+ },
+ {
+ "epoch": 12.14,
+ "learning_rate": 0.00030689102564102565,
+ "loss": 0.5311,
+ "step": 482
+ },
+ {
+ "epoch": 12.17,
+ "learning_rate": 0.00030649038461538465,
+ "loss": 0.5541,
+ "step": 483
+ },
+ {
+ "epoch": 12.2,
+ "learning_rate": 0.0003060897435897436,
+ "loss": 0.5242,
+ "step": 484
+ },
+ {
+ "epoch": 12.22,
+ "learning_rate": 0.0003056891025641026,
+ "loss": 0.5302,
+ "step": 485
+ },
+ {
+ "epoch": 12.25,
+ "learning_rate": 0.00030528846153846155,
+ "loss": 0.536,
+ "step": 486
+ },
+ {
+ "epoch": 12.27,
+ "learning_rate": 0.00030488782051282055,
+ "loss": 0.5545,
+ "step": 487
+ },
+ {
+ "epoch": 12.3,
+ "learning_rate": 0.00030448717948717945,
+ "loss": 0.5476,
+ "step": 488
+ },
+ {
+ "epoch": 12.32,
+ "learning_rate": 0.00030408653846153845,
+ "loss": 0.5528,
+ "step": 489
+ },
+ {
+ "epoch": 12.35,
+ "learning_rate": 0.0003036858974358974,
+ "loss": 0.5633,
+ "step": 490
+ },
+ {
+ "epoch": 12.37,
+ "learning_rate": 0.0003032852564102564,
+ "loss": 0.571,
+ "step": 491
+ },
+ {
+ "epoch": 12.4,
+ "learning_rate": 0.00030288461538461535,
+ "loss": 0.5637,
+ "step": 492
+ },
+ {
+ "epoch": 12.42,
+ "learning_rate": 0.00030248397435897435,
+ "loss": 0.5613,
+ "step": 493
+ },
+ {
+ "epoch": 12.45,
+ "learning_rate": 0.0003020833333333333,
+ "loss": 0.5999,
+ "step": 494
+ },
+ {
+ "epoch": 12.47,
+ "learning_rate": 0.0003016826923076923,
+ "loss": 0.5633,
+ "step": 495
+ },
+ {
+ "epoch": 12.5,
+ "learning_rate": 0.0003012820512820513,
+ "loss": 0.5811,
+ "step": 496
+ },
+ {
+ "epoch": 12.52,
+ "learning_rate": 0.00030088141025641025,
+ "loss": 0.5711,
+ "step": 497
+ },
+ {
+ "epoch": 12.55,
+ "learning_rate": 0.00030048076923076925,
+ "loss": 0.5669,
+ "step": 498
+ },
+ {
+ "epoch": 12.57,
+ "learning_rate": 0.0003000801282051282,
+ "loss": 0.5424,
+ "step": 499
+ },
+ {
+ "epoch": 12.6,
+ "learning_rate": 0.0002996794871794872,
+ "loss": 0.5592,
+ "step": 500
+ },
+ {
+ "epoch": 12.62,
+ "learning_rate": 0.00029927884615384614,
+ "loss": 0.5645,
+ "step": 501
+ },
+ {
+ "epoch": 12.65,
+ "learning_rate": 0.00029887820512820515,
+ "loss": 0.6089,
+ "step": 502
+ },
+ {
+ "epoch": 12.67,
+ "learning_rate": 0.0002984775641025641,
+ "loss": 0.57,
+ "step": 503
+ },
+ {
+ "epoch": 12.7,
+ "learning_rate": 0.0002980769230769231,
+ "loss": 0.57,
+ "step": 504
+ },
+ {
+ "epoch": 12.72,
+ "learning_rate": 0.00029767628205128204,
+ "loss": 0.5511,
+ "step": 505
+ },
+ {
+ "epoch": 12.75,
+ "learning_rate": 0.00029727564102564105,
+ "loss": 0.5765,
+ "step": 506
+ },
+ {
+ "epoch": 12.77,
+ "learning_rate": 0.000296875,
+ "loss": 0.5896,
+ "step": 507
+ },
+ {
+ "epoch": 12.8,
+ "learning_rate": 0.000296474358974359,
+ "loss": 0.6329,
+ "step": 508
+ },
+ {
+ "epoch": 12.83,
+ "learning_rate": 0.00029607371794871794,
+ "loss": 0.5682,
+ "step": 509
+ },
+ {
+ "epoch": 12.85,
+ "learning_rate": 0.00029567307692307694,
+ "loss": 0.58,
+ "step": 510
+ },
+ {
+ "epoch": 12.88,
+ "learning_rate": 0.0002952724358974359,
+ "loss": 0.6014,
+ "step": 511
+ },
+ {
+ "epoch": 12.9,
+ "learning_rate": 0.0002948717948717949,
+ "loss": 0.6082,
+ "step": 512
+ },
+ {
+ "epoch": 12.93,
+ "learning_rate": 0.00029447115384615384,
+ "loss": 0.6392,
+ "step": 513
+ },
+ {
+ "epoch": 12.95,
+ "learning_rate": 0.00029407051282051284,
+ "loss": 0.609,
+ "step": 514
+ },
+ {
+ "epoch": 12.98,
+ "learning_rate": 0.0002936698717948718,
+ "loss": 0.6247,
+ "step": 515
+ },
+ {
+ "epoch": 13.0,
+ "learning_rate": 0.0002932692307692308,
+ "loss": 0.5943,
+ "step": 516
+ },
+ {
+ "epoch": 13.03,
+ "learning_rate": 0.00029286858974358974,
+ "loss": 0.4499,
+ "step": 517
+ },
+ {
+ "epoch": 13.05,
+ "learning_rate": 0.00029246794871794874,
+ "loss": 0.4412,
+ "step": 518
+ },
+ {
+ "epoch": 13.08,
+ "learning_rate": 0.0002920673076923077,
+ "loss": 0.4409,
+ "step": 519
+ },
+ {
+ "epoch": 13.1,
+ "learning_rate": 0.0002916666666666667,
+ "loss": 0.4505,
+ "step": 520
+ },
+ {
+ "epoch": 13.13,
+ "learning_rate": 0.0002912660256410257,
+ "loss": 0.4489,
+ "step": 521
+ },
+ {
+ "epoch": 13.15,
+ "learning_rate": 0.00029086538461538464,
+ "loss": 0.4235,
+ "step": 522
+ },
+ {
+ "epoch": 13.18,
+ "learning_rate": 0.00029046474358974364,
+ "loss": 0.4276,
+ "step": 523
+ },
+ {
+ "epoch": 13.2,
+ "learning_rate": 0.0002900641025641026,
+ "loss": 0.4544,
+ "step": 524
+ },
+ {
+ "epoch": 13.23,
+ "learning_rate": 0.0002896634615384616,
+ "loss": 0.4625,
+ "step": 525
+ },
+ {
+ "epoch": 13.25,
+ "learning_rate": 0.00028926282051282054,
+ "loss": 0.4266,
+ "step": 526
+ },
+ {
+ "epoch": 13.28,
+ "learning_rate": 0.0002888621794871795,
+ "loss": 0.4459,
+ "step": 527
+ },
+ {
+ "epoch": 13.3,
+ "learning_rate": 0.00028846153846153843,
+ "loss": 0.4261,
+ "step": 528
+ },
+ {
+ "epoch": 13.33,
+ "learning_rate": 0.00028806089743589744,
+ "loss": 0.4376,
+ "step": 529
+ },
+ {
+ "epoch": 13.35,
+ "learning_rate": 0.0002876602564102564,
+ "loss": 0.4418,
+ "step": 530
+ },
+ {
+ "epoch": 13.38,
+ "learning_rate": 0.0002872596153846154,
+ "loss": 0.4667,
+ "step": 531
+ },
+ {
+ "epoch": 13.4,
+ "learning_rate": 0.00028685897435897433,
+ "loss": 0.4504,
+ "step": 532
+ },
+ {
+ "epoch": 13.43,
+ "learning_rate": 0.00028645833333333333,
+ "loss": 0.4691,
+ "step": 533
+ },
+ {
+ "epoch": 13.46,
+ "learning_rate": 0.0002860576923076923,
+ "loss": 0.4684,
+ "step": 534
+ },
+ {
+ "epoch": 13.48,
+ "learning_rate": 0.0002856570512820513,
+ "loss": 0.467,
+ "step": 535
+ },
+ {
+ "epoch": 13.51,
+ "learning_rate": 0.00028525641025641023,
+ "loss": 0.4716,
+ "step": 536
+ },
+ {
+ "epoch": 13.53,
+ "learning_rate": 0.00028485576923076923,
+ "loss": 0.4619,
+ "step": 537
+ },
+ {
+ "epoch": 13.56,
+ "learning_rate": 0.0002844551282051282,
+ "loss": 0.4775,
+ "step": 538
+ },
+ {
+ "epoch": 13.58,
+ "learning_rate": 0.0002840544871794872,
+ "loss": 0.4568,
+ "step": 539
+ },
+ {
+ "epoch": 13.61,
+ "learning_rate": 0.00028365384615384613,
+ "loss": 0.4763,
+ "step": 540
+ },
+ {
+ "epoch": 13.63,
+ "learning_rate": 0.00028325320512820513,
+ "loss": 0.466,
+ "step": 541
+ },
+ {
+ "epoch": 13.66,
+ "learning_rate": 0.0002828525641025641,
+ "loss": 0.4543,
+ "step": 542
+ },
+ {
+ "epoch": 13.68,
+ "learning_rate": 0.0002824519230769231,
+ "loss": 0.4666,
+ "step": 543
+ },
+ {
+ "epoch": 13.71,
+ "learning_rate": 0.00028205128205128203,
+ "loss": 0.489,
+ "step": 544
+ },
+ {
+ "epoch": 13.73,
+ "learning_rate": 0.00028165064102564103,
+ "loss": 0.491,
+ "step": 545
+ },
+ {
+ "epoch": 13.76,
+ "learning_rate": 0.00028125000000000003,
+ "loss": 0.4847,
+ "step": 546
+ },
+ {
+ "epoch": 13.78,
+ "learning_rate": 0.000280849358974359,
+ "loss": 0.4899,
+ "step": 547
+ },
+ {
+ "epoch": 13.81,
+ "learning_rate": 0.000280448717948718,
+ "loss": 0.5169,
+ "step": 548
+ },
+ {
+ "epoch": 13.83,
+ "learning_rate": 0.00028004807692307693,
+ "loss": 0.4968,
+ "step": 549
+ },
+ {
+ "epoch": 13.86,
+ "learning_rate": 0.00027964743589743593,
+ "loss": 0.4774,
+ "step": 550
+ },
+ {
+ "epoch": 13.88,
+ "learning_rate": 0.0002792467948717949,
+ "loss": 0.5105,
+ "step": 551
+ },
+ {
+ "epoch": 13.91,
+ "learning_rate": 0.0002788461538461539,
+ "loss": 0.4997,
+ "step": 552
+ },
+ {
+ "epoch": 13.93,
+ "learning_rate": 0.00027844551282051283,
+ "loss": 0.5088,
+ "step": 553
+ },
+ {
+ "epoch": 13.96,
+ "learning_rate": 0.00027804487179487183,
+ "loss": 0.5091,
+ "step": 554
+ },
+ {
+ "epoch": 13.98,
+ "learning_rate": 0.0002776442307692308,
+ "loss": 0.532,
+ "step": 555
+ },
+ {
+ "epoch": 14.01,
+ "learning_rate": 0.0002772435897435898,
+ "loss": 0.4487,
+ "step": 556
+ },
+ {
+ "epoch": 14.03,
+ "learning_rate": 0.00027684294871794873,
+ "loss": 0.3525,
+ "step": 557
+ },
+ {
+ "epoch": 14.06,
+ "learning_rate": 0.00027644230769230773,
+ "loss": 0.3715,
+ "step": 558
+ },
+ {
+ "epoch": 14.09,
+ "learning_rate": 0.0002760416666666667,
+ "loss": 0.3547,
+ "step": 559
+ },
+ {
+ "epoch": 14.11,
+ "learning_rate": 0.0002756410256410257,
+ "loss": 0.3862,
+ "step": 560
+ },
+ {
+ "epoch": 14.14,
+ "learning_rate": 0.0002752403846153846,
+ "loss": 0.3514,
+ "step": 561
+ },
+ {
+ "epoch": 14.16,
+ "learning_rate": 0.00027483974358974363,
+ "loss": 0.362,
+ "step": 562
+ },
+ {
+ "epoch": 14.19,
+ "learning_rate": 0.0002744391025641026,
+ "loss": 0.3457,
+ "step": 563
+ },
+ {
+ "epoch": 14.21,
+ "learning_rate": 0.0002740384615384616,
+ "loss": 0.3662,
+ "step": 564
+ },
+ {
+ "epoch": 14.24,
+ "learning_rate": 0.0002736378205128205,
+ "loss": 0.3746,
+ "step": 565
+ },
+ {
+ "epoch": 14.26,
+ "learning_rate": 0.00027323717948717947,
+ "loss": 0.3388,
+ "step": 566
+ },
+ {
+ "epoch": 14.29,
+ "learning_rate": 0.0002728365384615384,
+ "loss": 0.3824,
+ "step": 567
+ },
+ {
+ "epoch": 14.31,
+ "learning_rate": 0.0002724358974358974,
+ "loss": 0.3584,
+ "step": 568
+ },
+ {
+ "epoch": 14.34,
+ "learning_rate": 0.00027203525641025637,
+ "loss": 0.3544,
+ "step": 569
+ },
+ {
+ "epoch": 14.36,
+ "learning_rate": 0.00027163461538461537,
+ "loss": 0.3721,
+ "step": 570
+ },
+ {
+ "epoch": 14.39,
+ "learning_rate": 0.0002712339743589743,
+ "loss": 0.3769,
+ "step": 571
+ },
+ {
+ "epoch": 14.41,
+ "learning_rate": 0.0002708333333333333,
+ "loss": 0.3872,
+ "step": 572
+ },
+ {
+ "epoch": 14.44,
+ "learning_rate": 0.0002704326923076923,
+ "loss": 0.363,
+ "step": 573
+ },
+ {
+ "epoch": 14.46,
+ "learning_rate": 0.00027003205128205127,
+ "loss": 0.3842,
+ "step": 574
+ },
+ {
+ "epoch": 14.49,
+ "learning_rate": 0.00026963141025641027,
+ "loss": 0.3737,
+ "step": 575
+ },
+ {
+ "epoch": 14.51,
+ "learning_rate": 0.0002692307692307692,
+ "loss": 0.3981,
+ "step": 576
+ },
+ {
+ "epoch": 14.54,
+ "learning_rate": 0.0002688301282051282,
+ "loss": 0.3656,
+ "step": 577
+ },
+ {
+ "epoch": 14.56,
+ "learning_rate": 0.00026842948717948717,
+ "loss": 0.3852,
+ "step": 578
+ },
+ {
+ "epoch": 14.59,
+ "learning_rate": 0.00026802884615384617,
+ "loss": 0.3865,
+ "step": 579
+ },
+ {
+ "epoch": 14.61,
+ "learning_rate": 0.0002676282051282051,
+ "loss": 0.3747,
+ "step": 580
+ },
+ {
+ "epoch": 14.64,
+ "learning_rate": 0.0002672275641025641,
+ "loss": 0.3729,
+ "step": 581
+ },
+ {
+ "epoch": 14.66,
+ "learning_rate": 0.00026682692307692307,
+ "loss": 0.3785,
+ "step": 582
+ },
+ {
+ "epoch": 14.69,
+ "learning_rate": 0.00026642628205128207,
+ "loss": 0.4018,
+ "step": 583
+ },
+ {
+ "epoch": 14.71,
+ "learning_rate": 0.000266025641025641,
+ "loss": 0.3854,
+ "step": 584
+ },
+ {
+ "epoch": 14.74,
+ "learning_rate": 0.000265625,
+ "loss": 0.4161,
+ "step": 585
+ },
+ {
+ "epoch": 14.77,
+ "learning_rate": 0.00026522435897435897,
+ "loss": 0.4093,
+ "step": 586
+ },
+ {
+ "epoch": 14.79,
+ "learning_rate": 0.00026482371794871797,
+ "loss": 0.4126,
+ "step": 587
+ },
+ {
+ "epoch": 14.82,
+ "learning_rate": 0.0002644230769230769,
+ "loss": 0.3977,
+ "step": 588
+ },
+ {
+ "epoch": 14.84,
+ "learning_rate": 0.0002640224358974359,
+ "loss": 0.4286,
+ "step": 589
+ },
+ {
+ "epoch": 14.87,
+ "learning_rate": 0.00026362179487179487,
+ "loss": 0.4371,
+ "step": 590
+ },
+ {
+ "epoch": 14.89,
+ "learning_rate": 0.00026322115384615387,
+ "loss": 0.4066,
+ "step": 591
+ },
+ {
+ "epoch": 14.92,
+ "learning_rate": 0.0002628205128205128,
+ "loss": 0.433,
+ "step": 592
+ },
+ {
+ "epoch": 14.94,
+ "learning_rate": 0.0002624198717948718,
+ "loss": 0.4361,
+ "step": 593
+ },
+ {
+ "epoch": 14.97,
+ "learning_rate": 0.00026201923076923076,
+ "loss": 0.4276,
+ "step": 594
+ },
+ {
+ "epoch": 14.99,
+ "learning_rate": 0.00026161858974358977,
+ "loss": 0.4401,
+ "step": 595
+ },
+ {
+ "epoch": 15.02,
+ "learning_rate": 0.0002612179487179487,
+ "loss": 0.3514,
+ "step": 596
+ },
+ {
+ "epoch": 15.04,
+ "learning_rate": 0.0002608173076923077,
+ "loss": 0.306,
+ "step": 597
+ },
+ {
+ "epoch": 15.07,
+ "learning_rate": 0.0002604166666666667,
+ "loss": 0.2862,
+ "step": 598
+ },
+ {
+ "epoch": 15.09,
+ "learning_rate": 0.00026001602564102566,
+ "loss": 0.3042,
+ "step": 599
+ },
+ {
+ "epoch": 15.12,
+ "learning_rate": 0.00025961538461538467,
+ "loss": 0.2978,
+ "step": 600
+ },
+ {
+ "epoch": 15.14,
+ "learning_rate": 0.0002592147435897436,
+ "loss": 0.2975,
+ "step": 601
+ },
+ {
+ "epoch": 15.17,
+ "learning_rate": 0.0002588141025641026,
+ "loss": 0.3172,
+ "step": 602
+ },
+ {
+ "epoch": 15.19,
+ "learning_rate": 0.00025841346153846156,
+ "loss": 0.3046,
+ "step": 603
+ },
+ {
+ "epoch": 15.22,
+ "learning_rate": 0.00025801282051282057,
+ "loss": 0.3008,
+ "step": 604
+ },
+ {
+ "epoch": 15.24,
+ "learning_rate": 0.00025761217948717946,
+ "loss": 0.2956,
+ "step": 605
+ },
+ {
+ "epoch": 15.27,
+ "learning_rate": 0.00025721153846153846,
+ "loss": 0.3137,
+ "step": 606
+ },
+ {
+ "epoch": 15.29,
+ "learning_rate": 0.0002568108974358974,
+ "loss": 0.2961,
+ "step": 607
+ },
+ {
+ "epoch": 15.32,
+ "learning_rate": 0.0002564102564102564,
+ "loss": 0.3059,
+ "step": 608
+ },
+ {
+ "epoch": 15.34,
+ "learning_rate": 0.00025600961538461536,
+ "loss": 0.3077,
+ "step": 609
+ },
+ {
+ "epoch": 15.37,
+ "learning_rate": 0.00025560897435897436,
+ "loss": 0.3189,
+ "step": 610
+ },
+ {
+ "epoch": 15.4,
+ "learning_rate": 0.0002552083333333333,
+ "loss": 0.309,
+ "step": 611
+ },
+ {
+ "epoch": 15.42,
+ "learning_rate": 0.0002548076923076923,
+ "loss": 0.3324,
+ "step": 612
+ },
+ {
+ "epoch": 15.45,
+ "learning_rate": 0.00025440705128205126,
+ "loss": 0.3159,
+ "step": 613
+ },
+ {
+ "epoch": 15.47,
+ "learning_rate": 0.00025400641025641026,
+ "loss": 0.3188,
+ "step": 614
+ },
+ {
+ "epoch": 15.5,
+ "learning_rate": 0.0002536057692307692,
+ "loss": 0.3174,
+ "step": 615
+ },
+ {
+ "epoch": 15.52,
+ "learning_rate": 0.0002532051282051282,
+ "loss": 0.3074,
+ "step": 616
+ },
+ {
+ "epoch": 15.55,
+ "learning_rate": 0.00025280448717948716,
+ "loss": 0.3145,
+ "step": 617
+ },
+ {
+ "epoch": 15.57,
+ "learning_rate": 0.00025240384615384616,
+ "loss": 0.317,
+ "step": 618
+ },
+ {
+ "epoch": 15.6,
+ "learning_rate": 0.0002520032051282051,
+ "loss": 0.318,
+ "step": 619
+ },
+ {
+ "epoch": 15.62,
+ "learning_rate": 0.0002516025641025641,
+ "loss": 0.3236,
+ "step": 620
+ },
+ {
+ "epoch": 15.65,
+ "learning_rate": 0.00025120192307692305,
+ "loss": 0.3282,
+ "step": 621
+ },
+ {
+ "epoch": 15.67,
+ "learning_rate": 0.00025080128205128206,
+ "loss": 0.3394,
+ "step": 622
+ },
+ {
+ "epoch": 15.7,
+ "learning_rate": 0.000250400641025641,
+ "loss": 0.3278,
+ "step": 623
+ },
+ {
+ "epoch": 15.72,
+ "learning_rate": 0.00025,
+ "loss": 0.3255,
+ "step": 624
+ },
+ {
+ "epoch": 15.75,
+ "learning_rate": 0.000249599358974359,
+ "loss": 0.3256,
+ "step": 625
+ },
+ {
+ "epoch": 15.77,
+ "learning_rate": 0.00024919871794871795,
+ "loss": 0.3379,
+ "step": 626
+ },
+ {
+ "epoch": 15.8,
+ "learning_rate": 0.00024879807692307696,
+ "loss": 0.3384,
+ "step": 627
+ },
+ {
+ "epoch": 15.82,
+ "learning_rate": 0.0002483974358974359,
+ "loss": 0.3483,
+ "step": 628
+ },
+ {
+ "epoch": 15.85,
+ "learning_rate": 0.0002479967948717949,
+ "loss": 0.3478,
+ "step": 629
+ },
+ {
+ "epoch": 15.87,
+ "learning_rate": 0.00024759615384615385,
+ "loss": 0.337,
+ "step": 630
+ },
+ {
+ "epoch": 15.9,
+ "learning_rate": 0.00024719551282051286,
+ "loss": 0.3287,
+ "step": 631
+ },
+ {
+ "epoch": 15.92,
+ "learning_rate": 0.0002467948717948718,
+ "loss": 0.3478,
+ "step": 632
+ },
+ {
+ "epoch": 15.95,
+ "learning_rate": 0.0002463942307692308,
+ "loss": 0.3567,
+ "step": 633
+ },
+ {
+ "epoch": 15.97,
+ "learning_rate": 0.00024599358974358975,
+ "loss": 0.3359,
+ "step": 634
+ },
+ {
+ "epoch": 16.0,
+ "learning_rate": 0.0002455929487179487,
+ "loss": 0.3492,
+ "step": 635
+ },
+ {
+ "epoch": 16.03,
+ "learning_rate": 0.0002451923076923077,
+ "loss": 0.2497,
+ "step": 636
+ },
+ {
+ "epoch": 16.05,
+ "learning_rate": 0.00024479166666666665,
+ "loss": 0.2416,
+ "step": 637
+ },
+ {
+ "epoch": 16.08,
+ "learning_rate": 0.00024439102564102565,
+ "loss": 0.2576,
+ "step": 638
+ },
+ {
+ "epoch": 16.1,
+ "learning_rate": 0.00024399038461538463,
+ "loss": 0.243,
+ "step": 639
+ },
+ {
+ "epoch": 16.13,
+ "learning_rate": 0.0002435897435897436,
+ "loss": 0.2527,
+ "step": 640
+ },
+ {
+ "epoch": 16.15,
+ "learning_rate": 0.00024318910256410258,
+ "loss": 0.2606,
+ "step": 641
+ },
+ {
+ "epoch": 16.18,
+ "learning_rate": 0.00024278846153846155,
+ "loss": 0.2555,
+ "step": 642
+ },
+ {
+ "epoch": 16.2,
+ "learning_rate": 0.00024238782051282052,
+ "loss": 0.2581,
+ "step": 643
+ },
+ {
+ "epoch": 16.23,
+ "learning_rate": 0.0002419871794871795,
+ "loss": 0.2508,
+ "step": 644
+ },
+ {
+ "epoch": 16.25,
+ "learning_rate": 0.00024158653846153847,
+ "loss": 0.281,
+ "step": 645
+ },
+ {
+ "epoch": 16.28,
+ "learning_rate": 0.00024118589743589745,
+ "loss": 0.2626,
+ "step": 646
+ },
+ {
+ "epoch": 16.3,
+ "learning_rate": 0.00024078525641025642,
+ "loss": 0.2668,
+ "step": 647
+ },
+ {
+ "epoch": 16.33,
+ "learning_rate": 0.0002403846153846154,
+ "loss": 0.2675,
+ "step": 648
+ },
+ {
+ "epoch": 16.35,
+ "learning_rate": 0.00023998397435897437,
+ "loss": 0.2491,
+ "step": 649
+ },
+ {
+ "epoch": 16.38,
+ "learning_rate": 0.00023958333333333335,
+ "loss": 0.2459,
+ "step": 650
+ },
+ {
+ "epoch": 16.4,
+ "learning_rate": 0.00023918269230769232,
+ "loss": 0.2515,
+ "step": 651
+ },
+ {
+ "epoch": 16.43,
+ "learning_rate": 0.0002387820512820513,
+ "loss": 0.2612,
+ "step": 652
+ },
+ {
+ "epoch": 16.45,
+ "learning_rate": 0.00023838141025641027,
+ "loss": 0.2612,
+ "step": 653
+ },
+ {
+ "epoch": 16.48,
+ "learning_rate": 0.00023798076923076922,
+ "loss": 0.265,
+ "step": 654
+ },
+ {
+ "epoch": 16.5,
+ "learning_rate": 0.0002375801282051282,
+ "loss": 0.2634,
+ "step": 655
+ },
+ {
+ "epoch": 16.53,
+ "learning_rate": 0.00023717948717948717,
+ "loss": 0.2839,
+ "step": 656
+ },
+ {
+ "epoch": 16.55,
+ "learning_rate": 0.00023677884615384614,
+ "loss": 0.2595,
+ "step": 657
+ },
+ {
+ "epoch": 16.58,
+ "learning_rate": 0.00023637820512820512,
+ "loss": 0.2764,
+ "step": 658
+ },
+ {
+ "epoch": 16.6,
+ "learning_rate": 0.0002359775641025641,
+ "loss": 0.2553,
+ "step": 659
+ },
+ {
+ "epoch": 16.63,
+ "learning_rate": 0.00023557692307692307,
+ "loss": 0.2716,
+ "step": 660
+ },
+ {
+ "epoch": 16.66,
+ "learning_rate": 0.00023517628205128204,
+ "loss": 0.289,
+ "step": 661
+ },
+ {
+ "epoch": 16.68,
+ "learning_rate": 0.00023477564102564102,
+ "loss": 0.2483,
+ "step": 662
+ },
+ {
+ "epoch": 16.71,
+ "learning_rate": 0.000234375,
+ "loss": 0.269,
+ "step": 663
+ },
+ {
+ "epoch": 16.73,
+ "learning_rate": 0.000233974358974359,
+ "loss": 0.2738,
+ "step": 664
+ },
+ {
+ "epoch": 16.76,
+ "learning_rate": 0.00023357371794871797,
+ "loss": 0.2646,
+ "step": 665
+ },
+ {
+ "epoch": 16.78,
+ "learning_rate": 0.00023317307692307694,
+ "loss": 0.2793,
+ "step": 666
+ },
+ {
+ "epoch": 16.81,
+ "learning_rate": 0.00023277243589743592,
+ "loss": 0.2666,
+ "step": 667
+ },
+ {
+ "epoch": 16.83,
+ "learning_rate": 0.0002323717948717949,
+ "loss": 0.2799,
+ "step": 668
+ },
+ {
+ "epoch": 16.86,
+ "learning_rate": 0.00023197115384615387,
+ "loss": 0.2677,
+ "step": 669
+ },
+ {
+ "epoch": 16.88,
+ "learning_rate": 0.00023157051282051284,
+ "loss": 0.2889,
+ "step": 670
+ },
+ {
+ "epoch": 16.91,
+ "learning_rate": 0.00023116987179487182,
+ "loss": 0.2869,
+ "step": 671
+ },
+ {
+ "epoch": 16.93,
+ "learning_rate": 0.0002307692307692308,
+ "loss": 0.2814,
+ "step": 672
+ },
+ {
+ "epoch": 16.96,
+ "learning_rate": 0.00023036858974358974,
+ "loss": 0.2881,
+ "step": 673
+ },
+ {
+ "epoch": 16.98,
+ "learning_rate": 0.0002299679487179487,
+ "loss": 0.2777,
+ "step": 674
+ },
+ {
+ "epoch": 17.01,
+ "learning_rate": 0.0002295673076923077,
+ "loss": 0.2713,
+ "step": 675
+ },
+ {
+ "epoch": 17.03,
+ "learning_rate": 0.00022916666666666666,
+ "loss": 0.2176,
+ "step": 676
+ },
+ {
+ "epoch": 17.06,
+ "learning_rate": 0.00022876602564102564,
+ "loss": 0.2032,
+ "step": 677
+ },
+ {
+ "epoch": 17.08,
+ "learning_rate": 0.0002283653846153846,
+ "loss": 0.2207,
+ "step": 678
+ },
+ {
+ "epoch": 17.11,
+ "learning_rate": 0.0002279647435897436,
+ "loss": 0.2183,
+ "step": 679
+ },
+ {
+ "epoch": 17.13,
+ "learning_rate": 0.00022756410256410256,
+ "loss": 0.2238,
+ "step": 680
+ },
+ {
+ "epoch": 17.16,
+ "learning_rate": 0.00022716346153846154,
+ "loss": 0.2126,
+ "step": 681
+ },
+ {
+ "epoch": 17.18,
+ "learning_rate": 0.0002267628205128205,
+ "loss": 0.2104,
+ "step": 682
+ },
+ {
+ "epoch": 17.21,
+ "learning_rate": 0.00022636217948717949,
+ "loss": 0.2151,
+ "step": 683
+ },
+ {
+ "epoch": 17.23,
+ "learning_rate": 0.00022596153846153846,
+ "loss": 0.2093,
+ "step": 684
+ },
+ {
+ "epoch": 17.26,
+ "learning_rate": 0.00022556089743589743,
+ "loss": 0.2308,
+ "step": 685
+ },
+ {
+ "epoch": 17.29,
+ "learning_rate": 0.0002251602564102564,
+ "loss": 0.2276,
+ "step": 686
+ },
+ {
+ "epoch": 17.31,
+ "learning_rate": 0.00022475961538461538,
+ "loss": 0.2328,
+ "step": 687
+ },
+ {
+ "epoch": 17.34,
+ "learning_rate": 0.00022435897435897436,
+ "loss": 0.2076,
+ "step": 688
+ },
+ {
+ "epoch": 17.36,
+ "learning_rate": 0.00022395833333333336,
+ "loss": 0.2125,
+ "step": 689
+ },
+ {
+ "epoch": 17.39,
+ "learning_rate": 0.00022355769230769234,
+ "loss": 0.2181,
+ "step": 690
+ },
+ {
+ "epoch": 17.41,
+ "learning_rate": 0.0002231570512820513,
+ "loss": 0.2153,
+ "step": 691
+ },
+ {
+ "epoch": 17.44,
+ "learning_rate": 0.00022275641025641028,
+ "loss": 0.2122,
+ "step": 692
+ },
+ {
+ "epoch": 17.46,
+ "learning_rate": 0.00022235576923076923,
+ "loss": 0.2262,
+ "step": 693
+ },
+ {
+ "epoch": 17.49,
+ "learning_rate": 0.0002219551282051282,
+ "loss": 0.2206,
+ "step": 694
+ },
+ {
+ "epoch": 17.51,
+ "learning_rate": 0.00022155448717948718,
+ "loss": 0.2223,
+ "step": 695
+ },
+ {
+ "epoch": 17.54,
+ "learning_rate": 0.00022115384615384616,
+ "loss": 0.2306,
+ "step": 696
+ },
+ {
+ "epoch": 17.56,
+ "learning_rate": 0.00022075320512820513,
+ "loss": 0.213,
+ "step": 697
+ },
+ {
+ "epoch": 17.59,
+ "learning_rate": 0.0002203525641025641,
+ "loss": 0.2198,
+ "step": 698
+ },
+ {
+ "epoch": 17.61,
+ "learning_rate": 0.00021995192307692308,
+ "loss": 0.2299,
+ "step": 699
+ },
+ {
+ "epoch": 17.64,
+ "learning_rate": 0.00021955128205128206,
+ "loss": 0.2213,
+ "step": 700
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 3.935420097143808e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-700/training_args.bin b/checkpoint-700/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-700/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-800/README.md b/checkpoint-800/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-800/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-800/adapter_config.json b/checkpoint-800/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-800/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-800/adapter_model.bin b/checkpoint-800/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..64025b74c21ec428befef748f450f3279ab1aa92
--- /dev/null
+++ b/checkpoint-800/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1de9c3b542c8ebb8e945732fdd9cf34f75518fed71fa667262ac263ae280674a
+size 39409357
diff --git a/checkpoint-800/optimizer.pt b/checkpoint-800/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b531643d3073c05cf35e3fb626dbdda3be55e873
--- /dev/null
+++ b/checkpoint-800/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:465ba7e77006068fefb8ac3155fb4fb88b9edd933be5fa4a006146ca2b152976
+size 78844421
diff --git a/checkpoint-800/rng_state.pth b/checkpoint-800/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..faca5241da3506eadd57588fb689a90f390f5558
--- /dev/null
+++ b/checkpoint-800/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a5432f961e7c82cbe7b6ec30d027cf3db4d17372f65774be8cf037a208cbc1ac
+size 14575
diff --git a/checkpoint-800/scheduler.pt b/checkpoint-800/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..30c8b87700d57c9a65be0cfef4a20b2fb4a99474
--- /dev/null
+++ b/checkpoint-800/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2af5cf7381b6a582b236b6d483df3750dec8b41d92fdf59ebee031b0a2374146
+size 627
diff --git a/checkpoint-800/special_tokens_map.json b/checkpoint-800/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-800/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-800/tokenizer.json b/checkpoint-800/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-800/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-800/tokenizer_config.json b/checkpoint-800/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-800/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-800/trainer_state.json b/checkpoint-800/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..61d059bf8d86dd32ebe29d4907c3c426f0e5505d
--- /dev/null
+++ b/checkpoint-800/trainer_state.json
@@ -0,0 +1,4819 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 20.15748031496063,
+ "eval_steps": 500,
+ "global_step": 800,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ },
+ {
+ "epoch": 10.1,
+ "learning_rate": 0.00033934294871794873,
+ "loss": 0.7635,
+ "step": 401
+ },
+ {
+ "epoch": 10.13,
+ "learning_rate": 0.00033894230769230773,
+ "loss": 0.7713,
+ "step": 402
+ },
+ {
+ "epoch": 10.15,
+ "learning_rate": 0.0003385416666666667,
+ "loss": 0.827,
+ "step": 403
+ },
+ {
+ "epoch": 10.18,
+ "learning_rate": 0.0003381410256410257,
+ "loss": 0.7844,
+ "step": 404
+ },
+ {
+ "epoch": 10.2,
+ "learning_rate": 0.00033774038461538463,
+ "loss": 0.7839,
+ "step": 405
+ },
+ {
+ "epoch": 10.23,
+ "learning_rate": 0.00033733974358974363,
+ "loss": 0.7702,
+ "step": 406
+ },
+ {
+ "epoch": 10.26,
+ "learning_rate": 0.0003369391025641026,
+ "loss": 0.7855,
+ "step": 407
+ },
+ {
+ "epoch": 10.28,
+ "learning_rate": 0.0003365384615384616,
+ "loss": 0.7669,
+ "step": 408
+ },
+ {
+ "epoch": 10.31,
+ "learning_rate": 0.0003361378205128205,
+ "loss": 0.8054,
+ "step": 409
+ },
+ {
+ "epoch": 10.33,
+ "learning_rate": 0.0003357371794871795,
+ "loss": 0.7792,
+ "step": 410
+ },
+ {
+ "epoch": 10.36,
+ "learning_rate": 0.0003353365384615384,
+ "loss": 0.7986,
+ "step": 411
+ },
+ {
+ "epoch": 10.38,
+ "learning_rate": 0.0003349358974358974,
+ "loss": 0.8269,
+ "step": 412
+ },
+ {
+ "epoch": 10.41,
+ "learning_rate": 0.00033453525641025637,
+ "loss": 0.7923,
+ "step": 413
+ },
+ {
+ "epoch": 10.43,
+ "learning_rate": 0.00033413461538461537,
+ "loss": 0.8242,
+ "step": 414
+ },
+ {
+ "epoch": 10.46,
+ "learning_rate": 0.0003337339743589743,
+ "loss": 0.8496,
+ "step": 415
+ },
+ {
+ "epoch": 10.48,
+ "learning_rate": 0.0003333333333333333,
+ "loss": 0.8236,
+ "step": 416
+ },
+ {
+ "epoch": 10.51,
+ "learning_rate": 0.0003329326923076923,
+ "loss": 0.8685,
+ "step": 417
+ },
+ {
+ "epoch": 10.53,
+ "learning_rate": 0.00033253205128205127,
+ "loss": 0.8542,
+ "step": 418
+ },
+ {
+ "epoch": 10.56,
+ "learning_rate": 0.0003321314102564103,
+ "loss": 0.8185,
+ "step": 419
+ },
+ {
+ "epoch": 10.58,
+ "learning_rate": 0.0003317307692307692,
+ "loss": 0.8477,
+ "step": 420
+ },
+ {
+ "epoch": 10.61,
+ "learning_rate": 0.0003313301282051282,
+ "loss": 0.8372,
+ "step": 421
+ },
+ {
+ "epoch": 10.63,
+ "learning_rate": 0.00033092948717948717,
+ "loss": 0.8731,
+ "step": 422
+ },
+ {
+ "epoch": 10.66,
+ "learning_rate": 0.00033052884615384617,
+ "loss": 0.8322,
+ "step": 423
+ },
+ {
+ "epoch": 10.68,
+ "learning_rate": 0.0003301282051282051,
+ "loss": 0.8514,
+ "step": 424
+ },
+ {
+ "epoch": 10.71,
+ "learning_rate": 0.0003297275641025641,
+ "loss": 0.8564,
+ "step": 425
+ },
+ {
+ "epoch": 10.73,
+ "learning_rate": 0.00032932692307692307,
+ "loss": 0.8278,
+ "step": 426
+ },
+ {
+ "epoch": 10.76,
+ "learning_rate": 0.00032892628205128207,
+ "loss": 0.8408,
+ "step": 427
+ },
+ {
+ "epoch": 10.78,
+ "learning_rate": 0.000328525641025641,
+ "loss": 0.8438,
+ "step": 428
+ },
+ {
+ "epoch": 10.81,
+ "learning_rate": 0.000328125,
+ "loss": 0.8862,
+ "step": 429
+ },
+ {
+ "epoch": 10.83,
+ "learning_rate": 0.00032772435897435897,
+ "loss": 0.8631,
+ "step": 430
+ },
+ {
+ "epoch": 10.86,
+ "learning_rate": 0.00032732371794871797,
+ "loss": 0.883,
+ "step": 431
+ },
+ {
+ "epoch": 10.89,
+ "learning_rate": 0.0003269230769230769,
+ "loss": 0.8644,
+ "step": 432
+ },
+ {
+ "epoch": 10.91,
+ "learning_rate": 0.0003265224358974359,
+ "loss": 0.8907,
+ "step": 433
+ },
+ {
+ "epoch": 10.94,
+ "learning_rate": 0.00032612179487179487,
+ "loss": 0.9141,
+ "step": 434
+ },
+ {
+ "epoch": 10.96,
+ "learning_rate": 0.00032572115384615387,
+ "loss": 0.8877,
+ "step": 435
+ },
+ {
+ "epoch": 10.99,
+ "learning_rate": 0.0003253205128205128,
+ "loss": 0.8521,
+ "step": 436
+ },
+ {
+ "epoch": 11.01,
+ "learning_rate": 0.0003249198717948718,
+ "loss": 0.7842,
+ "step": 437
+ },
+ {
+ "epoch": 11.04,
+ "learning_rate": 0.00032451923076923077,
+ "loss": 0.6516,
+ "step": 438
+ },
+ {
+ "epoch": 11.06,
+ "learning_rate": 0.00032411858974358977,
+ "loss": 0.6549,
+ "step": 439
+ },
+ {
+ "epoch": 11.09,
+ "learning_rate": 0.0003237179487179487,
+ "loss": 0.6736,
+ "step": 440
+ },
+ {
+ "epoch": 11.11,
+ "learning_rate": 0.0003233173076923077,
+ "loss": 0.6625,
+ "step": 441
+ },
+ {
+ "epoch": 11.14,
+ "learning_rate": 0.0003229166666666667,
+ "loss": 0.6466,
+ "step": 442
+ },
+ {
+ "epoch": 11.16,
+ "learning_rate": 0.00032251602564102567,
+ "loss": 0.6362,
+ "step": 443
+ },
+ {
+ "epoch": 11.19,
+ "learning_rate": 0.00032211538461538467,
+ "loss": 0.674,
+ "step": 444
+ },
+ {
+ "epoch": 11.21,
+ "learning_rate": 0.0003217147435897436,
+ "loss": 0.6509,
+ "step": 445
+ },
+ {
+ "epoch": 11.24,
+ "learning_rate": 0.0003213141025641026,
+ "loss": 0.618,
+ "step": 446
+ },
+ {
+ "epoch": 11.26,
+ "learning_rate": 0.00032091346153846156,
+ "loss": 0.6654,
+ "step": 447
+ },
+ {
+ "epoch": 11.29,
+ "learning_rate": 0.00032051282051282057,
+ "loss": 0.6632,
+ "step": 448
+ },
+ {
+ "epoch": 11.31,
+ "learning_rate": 0.00032011217948717946,
+ "loss": 0.6599,
+ "step": 449
+ },
+ {
+ "epoch": 11.34,
+ "learning_rate": 0.00031971153846153846,
+ "loss": 0.6537,
+ "step": 450
+ },
+ {
+ "epoch": 11.36,
+ "learning_rate": 0.0003193108974358974,
+ "loss": 0.678,
+ "step": 451
+ },
+ {
+ "epoch": 11.39,
+ "learning_rate": 0.0003189102564102564,
+ "loss": 0.684,
+ "step": 452
+ },
+ {
+ "epoch": 11.41,
+ "learning_rate": 0.00031850961538461536,
+ "loss": 0.6419,
+ "step": 453
+ },
+ {
+ "epoch": 11.44,
+ "learning_rate": 0.00031810897435897436,
+ "loss": 0.6614,
+ "step": 454
+ },
+ {
+ "epoch": 11.46,
+ "learning_rate": 0.0003177083333333333,
+ "loss": 0.7037,
+ "step": 455
+ },
+ {
+ "epoch": 11.49,
+ "learning_rate": 0.0003173076923076923,
+ "loss": 0.6944,
+ "step": 456
+ },
+ {
+ "epoch": 11.51,
+ "learning_rate": 0.00031690705128205126,
+ "loss": 0.6818,
+ "step": 457
+ },
+ {
+ "epoch": 11.54,
+ "learning_rate": 0.00031650641025641026,
+ "loss": 0.692,
+ "step": 458
+ },
+ {
+ "epoch": 11.57,
+ "learning_rate": 0.0003161057692307692,
+ "loss": 0.7028,
+ "step": 459
+ },
+ {
+ "epoch": 11.59,
+ "learning_rate": 0.0003157051282051282,
+ "loss": 0.7009,
+ "step": 460
+ },
+ {
+ "epoch": 11.62,
+ "learning_rate": 0.00031530448717948716,
+ "loss": 0.7325,
+ "step": 461
+ },
+ {
+ "epoch": 11.64,
+ "learning_rate": 0.00031490384615384616,
+ "loss": 0.6738,
+ "step": 462
+ },
+ {
+ "epoch": 11.67,
+ "learning_rate": 0.0003145032051282051,
+ "loss": 0.7008,
+ "step": 463
+ },
+ {
+ "epoch": 11.69,
+ "learning_rate": 0.0003141025641025641,
+ "loss": 0.6838,
+ "step": 464
+ },
+ {
+ "epoch": 11.72,
+ "learning_rate": 0.00031370192307692306,
+ "loss": 0.726,
+ "step": 465
+ },
+ {
+ "epoch": 11.74,
+ "learning_rate": 0.00031330128205128206,
+ "loss": 0.7167,
+ "step": 466
+ },
+ {
+ "epoch": 11.77,
+ "learning_rate": 0.000312900641025641,
+ "loss": 0.7236,
+ "step": 467
+ },
+ {
+ "epoch": 11.79,
+ "learning_rate": 0.0003125,
+ "loss": 0.7076,
+ "step": 468
+ },
+ {
+ "epoch": 11.82,
+ "learning_rate": 0.000312099358974359,
+ "loss": 0.7415,
+ "step": 469
+ },
+ {
+ "epoch": 11.84,
+ "learning_rate": 0.00031169871794871796,
+ "loss": 0.7313,
+ "step": 470
+ },
+ {
+ "epoch": 11.87,
+ "learning_rate": 0.00031129807692307696,
+ "loss": 0.7403,
+ "step": 471
+ },
+ {
+ "epoch": 11.89,
+ "learning_rate": 0.0003108974358974359,
+ "loss": 0.7145,
+ "step": 472
+ },
+ {
+ "epoch": 11.92,
+ "learning_rate": 0.0003104967948717949,
+ "loss": 0.7532,
+ "step": 473
+ },
+ {
+ "epoch": 11.94,
+ "learning_rate": 0.00031009615384615385,
+ "loss": 0.7326,
+ "step": 474
+ },
+ {
+ "epoch": 11.97,
+ "learning_rate": 0.00030969551282051286,
+ "loss": 0.7699,
+ "step": 475
+ },
+ {
+ "epoch": 11.99,
+ "learning_rate": 0.0003092948717948718,
+ "loss": 0.7251,
+ "step": 476
+ },
+ {
+ "epoch": 12.02,
+ "learning_rate": 0.0003088942307692308,
+ "loss": 0.5773,
+ "step": 477
+ },
+ {
+ "epoch": 12.04,
+ "learning_rate": 0.00030849358974358975,
+ "loss": 0.5288,
+ "step": 478
+ },
+ {
+ "epoch": 12.07,
+ "learning_rate": 0.00030809294871794876,
+ "loss": 0.5312,
+ "step": 479
+ },
+ {
+ "epoch": 12.09,
+ "learning_rate": 0.0003076923076923077,
+ "loss": 0.5445,
+ "step": 480
+ },
+ {
+ "epoch": 12.12,
+ "learning_rate": 0.0003072916666666667,
+ "loss": 0.5218,
+ "step": 481
+ },
+ {
+ "epoch": 12.14,
+ "learning_rate": 0.00030689102564102565,
+ "loss": 0.5311,
+ "step": 482
+ },
+ {
+ "epoch": 12.17,
+ "learning_rate": 0.00030649038461538465,
+ "loss": 0.5541,
+ "step": 483
+ },
+ {
+ "epoch": 12.2,
+ "learning_rate": 0.0003060897435897436,
+ "loss": 0.5242,
+ "step": 484
+ },
+ {
+ "epoch": 12.22,
+ "learning_rate": 0.0003056891025641026,
+ "loss": 0.5302,
+ "step": 485
+ },
+ {
+ "epoch": 12.25,
+ "learning_rate": 0.00030528846153846155,
+ "loss": 0.536,
+ "step": 486
+ },
+ {
+ "epoch": 12.27,
+ "learning_rate": 0.00030488782051282055,
+ "loss": 0.5545,
+ "step": 487
+ },
+ {
+ "epoch": 12.3,
+ "learning_rate": 0.00030448717948717945,
+ "loss": 0.5476,
+ "step": 488
+ },
+ {
+ "epoch": 12.32,
+ "learning_rate": 0.00030408653846153845,
+ "loss": 0.5528,
+ "step": 489
+ },
+ {
+ "epoch": 12.35,
+ "learning_rate": 0.0003036858974358974,
+ "loss": 0.5633,
+ "step": 490
+ },
+ {
+ "epoch": 12.37,
+ "learning_rate": 0.0003032852564102564,
+ "loss": 0.571,
+ "step": 491
+ },
+ {
+ "epoch": 12.4,
+ "learning_rate": 0.00030288461538461535,
+ "loss": 0.5637,
+ "step": 492
+ },
+ {
+ "epoch": 12.42,
+ "learning_rate": 0.00030248397435897435,
+ "loss": 0.5613,
+ "step": 493
+ },
+ {
+ "epoch": 12.45,
+ "learning_rate": 0.0003020833333333333,
+ "loss": 0.5999,
+ "step": 494
+ },
+ {
+ "epoch": 12.47,
+ "learning_rate": 0.0003016826923076923,
+ "loss": 0.5633,
+ "step": 495
+ },
+ {
+ "epoch": 12.5,
+ "learning_rate": 0.0003012820512820513,
+ "loss": 0.5811,
+ "step": 496
+ },
+ {
+ "epoch": 12.52,
+ "learning_rate": 0.00030088141025641025,
+ "loss": 0.5711,
+ "step": 497
+ },
+ {
+ "epoch": 12.55,
+ "learning_rate": 0.00030048076923076925,
+ "loss": 0.5669,
+ "step": 498
+ },
+ {
+ "epoch": 12.57,
+ "learning_rate": 0.0003000801282051282,
+ "loss": 0.5424,
+ "step": 499
+ },
+ {
+ "epoch": 12.6,
+ "learning_rate": 0.0002996794871794872,
+ "loss": 0.5592,
+ "step": 500
+ },
+ {
+ "epoch": 12.62,
+ "learning_rate": 0.00029927884615384614,
+ "loss": 0.5645,
+ "step": 501
+ },
+ {
+ "epoch": 12.65,
+ "learning_rate": 0.00029887820512820515,
+ "loss": 0.6089,
+ "step": 502
+ },
+ {
+ "epoch": 12.67,
+ "learning_rate": 0.0002984775641025641,
+ "loss": 0.57,
+ "step": 503
+ },
+ {
+ "epoch": 12.7,
+ "learning_rate": 0.0002980769230769231,
+ "loss": 0.57,
+ "step": 504
+ },
+ {
+ "epoch": 12.72,
+ "learning_rate": 0.00029767628205128204,
+ "loss": 0.5511,
+ "step": 505
+ },
+ {
+ "epoch": 12.75,
+ "learning_rate": 0.00029727564102564105,
+ "loss": 0.5765,
+ "step": 506
+ },
+ {
+ "epoch": 12.77,
+ "learning_rate": 0.000296875,
+ "loss": 0.5896,
+ "step": 507
+ },
+ {
+ "epoch": 12.8,
+ "learning_rate": 0.000296474358974359,
+ "loss": 0.6329,
+ "step": 508
+ },
+ {
+ "epoch": 12.83,
+ "learning_rate": 0.00029607371794871794,
+ "loss": 0.5682,
+ "step": 509
+ },
+ {
+ "epoch": 12.85,
+ "learning_rate": 0.00029567307692307694,
+ "loss": 0.58,
+ "step": 510
+ },
+ {
+ "epoch": 12.88,
+ "learning_rate": 0.0002952724358974359,
+ "loss": 0.6014,
+ "step": 511
+ },
+ {
+ "epoch": 12.9,
+ "learning_rate": 0.0002948717948717949,
+ "loss": 0.6082,
+ "step": 512
+ },
+ {
+ "epoch": 12.93,
+ "learning_rate": 0.00029447115384615384,
+ "loss": 0.6392,
+ "step": 513
+ },
+ {
+ "epoch": 12.95,
+ "learning_rate": 0.00029407051282051284,
+ "loss": 0.609,
+ "step": 514
+ },
+ {
+ "epoch": 12.98,
+ "learning_rate": 0.0002936698717948718,
+ "loss": 0.6247,
+ "step": 515
+ },
+ {
+ "epoch": 13.0,
+ "learning_rate": 0.0002932692307692308,
+ "loss": 0.5943,
+ "step": 516
+ },
+ {
+ "epoch": 13.03,
+ "learning_rate": 0.00029286858974358974,
+ "loss": 0.4499,
+ "step": 517
+ },
+ {
+ "epoch": 13.05,
+ "learning_rate": 0.00029246794871794874,
+ "loss": 0.4412,
+ "step": 518
+ },
+ {
+ "epoch": 13.08,
+ "learning_rate": 0.0002920673076923077,
+ "loss": 0.4409,
+ "step": 519
+ },
+ {
+ "epoch": 13.1,
+ "learning_rate": 0.0002916666666666667,
+ "loss": 0.4505,
+ "step": 520
+ },
+ {
+ "epoch": 13.13,
+ "learning_rate": 0.0002912660256410257,
+ "loss": 0.4489,
+ "step": 521
+ },
+ {
+ "epoch": 13.15,
+ "learning_rate": 0.00029086538461538464,
+ "loss": 0.4235,
+ "step": 522
+ },
+ {
+ "epoch": 13.18,
+ "learning_rate": 0.00029046474358974364,
+ "loss": 0.4276,
+ "step": 523
+ },
+ {
+ "epoch": 13.2,
+ "learning_rate": 0.0002900641025641026,
+ "loss": 0.4544,
+ "step": 524
+ },
+ {
+ "epoch": 13.23,
+ "learning_rate": 0.0002896634615384616,
+ "loss": 0.4625,
+ "step": 525
+ },
+ {
+ "epoch": 13.25,
+ "learning_rate": 0.00028926282051282054,
+ "loss": 0.4266,
+ "step": 526
+ },
+ {
+ "epoch": 13.28,
+ "learning_rate": 0.0002888621794871795,
+ "loss": 0.4459,
+ "step": 527
+ },
+ {
+ "epoch": 13.3,
+ "learning_rate": 0.00028846153846153843,
+ "loss": 0.4261,
+ "step": 528
+ },
+ {
+ "epoch": 13.33,
+ "learning_rate": 0.00028806089743589744,
+ "loss": 0.4376,
+ "step": 529
+ },
+ {
+ "epoch": 13.35,
+ "learning_rate": 0.0002876602564102564,
+ "loss": 0.4418,
+ "step": 530
+ },
+ {
+ "epoch": 13.38,
+ "learning_rate": 0.0002872596153846154,
+ "loss": 0.4667,
+ "step": 531
+ },
+ {
+ "epoch": 13.4,
+ "learning_rate": 0.00028685897435897433,
+ "loss": 0.4504,
+ "step": 532
+ },
+ {
+ "epoch": 13.43,
+ "learning_rate": 0.00028645833333333333,
+ "loss": 0.4691,
+ "step": 533
+ },
+ {
+ "epoch": 13.46,
+ "learning_rate": 0.0002860576923076923,
+ "loss": 0.4684,
+ "step": 534
+ },
+ {
+ "epoch": 13.48,
+ "learning_rate": 0.0002856570512820513,
+ "loss": 0.467,
+ "step": 535
+ },
+ {
+ "epoch": 13.51,
+ "learning_rate": 0.00028525641025641023,
+ "loss": 0.4716,
+ "step": 536
+ },
+ {
+ "epoch": 13.53,
+ "learning_rate": 0.00028485576923076923,
+ "loss": 0.4619,
+ "step": 537
+ },
+ {
+ "epoch": 13.56,
+ "learning_rate": 0.0002844551282051282,
+ "loss": 0.4775,
+ "step": 538
+ },
+ {
+ "epoch": 13.58,
+ "learning_rate": 0.0002840544871794872,
+ "loss": 0.4568,
+ "step": 539
+ },
+ {
+ "epoch": 13.61,
+ "learning_rate": 0.00028365384615384613,
+ "loss": 0.4763,
+ "step": 540
+ },
+ {
+ "epoch": 13.63,
+ "learning_rate": 0.00028325320512820513,
+ "loss": 0.466,
+ "step": 541
+ },
+ {
+ "epoch": 13.66,
+ "learning_rate": 0.0002828525641025641,
+ "loss": 0.4543,
+ "step": 542
+ },
+ {
+ "epoch": 13.68,
+ "learning_rate": 0.0002824519230769231,
+ "loss": 0.4666,
+ "step": 543
+ },
+ {
+ "epoch": 13.71,
+ "learning_rate": 0.00028205128205128203,
+ "loss": 0.489,
+ "step": 544
+ },
+ {
+ "epoch": 13.73,
+ "learning_rate": 0.00028165064102564103,
+ "loss": 0.491,
+ "step": 545
+ },
+ {
+ "epoch": 13.76,
+ "learning_rate": 0.00028125000000000003,
+ "loss": 0.4847,
+ "step": 546
+ },
+ {
+ "epoch": 13.78,
+ "learning_rate": 0.000280849358974359,
+ "loss": 0.4899,
+ "step": 547
+ },
+ {
+ "epoch": 13.81,
+ "learning_rate": 0.000280448717948718,
+ "loss": 0.5169,
+ "step": 548
+ },
+ {
+ "epoch": 13.83,
+ "learning_rate": 0.00028004807692307693,
+ "loss": 0.4968,
+ "step": 549
+ },
+ {
+ "epoch": 13.86,
+ "learning_rate": 0.00027964743589743593,
+ "loss": 0.4774,
+ "step": 550
+ },
+ {
+ "epoch": 13.88,
+ "learning_rate": 0.0002792467948717949,
+ "loss": 0.5105,
+ "step": 551
+ },
+ {
+ "epoch": 13.91,
+ "learning_rate": 0.0002788461538461539,
+ "loss": 0.4997,
+ "step": 552
+ },
+ {
+ "epoch": 13.93,
+ "learning_rate": 0.00027844551282051283,
+ "loss": 0.5088,
+ "step": 553
+ },
+ {
+ "epoch": 13.96,
+ "learning_rate": 0.00027804487179487183,
+ "loss": 0.5091,
+ "step": 554
+ },
+ {
+ "epoch": 13.98,
+ "learning_rate": 0.0002776442307692308,
+ "loss": 0.532,
+ "step": 555
+ },
+ {
+ "epoch": 14.01,
+ "learning_rate": 0.0002772435897435898,
+ "loss": 0.4487,
+ "step": 556
+ },
+ {
+ "epoch": 14.03,
+ "learning_rate": 0.00027684294871794873,
+ "loss": 0.3525,
+ "step": 557
+ },
+ {
+ "epoch": 14.06,
+ "learning_rate": 0.00027644230769230773,
+ "loss": 0.3715,
+ "step": 558
+ },
+ {
+ "epoch": 14.09,
+ "learning_rate": 0.0002760416666666667,
+ "loss": 0.3547,
+ "step": 559
+ },
+ {
+ "epoch": 14.11,
+ "learning_rate": 0.0002756410256410257,
+ "loss": 0.3862,
+ "step": 560
+ },
+ {
+ "epoch": 14.14,
+ "learning_rate": 0.0002752403846153846,
+ "loss": 0.3514,
+ "step": 561
+ },
+ {
+ "epoch": 14.16,
+ "learning_rate": 0.00027483974358974363,
+ "loss": 0.362,
+ "step": 562
+ },
+ {
+ "epoch": 14.19,
+ "learning_rate": 0.0002744391025641026,
+ "loss": 0.3457,
+ "step": 563
+ },
+ {
+ "epoch": 14.21,
+ "learning_rate": 0.0002740384615384616,
+ "loss": 0.3662,
+ "step": 564
+ },
+ {
+ "epoch": 14.24,
+ "learning_rate": 0.0002736378205128205,
+ "loss": 0.3746,
+ "step": 565
+ },
+ {
+ "epoch": 14.26,
+ "learning_rate": 0.00027323717948717947,
+ "loss": 0.3388,
+ "step": 566
+ },
+ {
+ "epoch": 14.29,
+ "learning_rate": 0.0002728365384615384,
+ "loss": 0.3824,
+ "step": 567
+ },
+ {
+ "epoch": 14.31,
+ "learning_rate": 0.0002724358974358974,
+ "loss": 0.3584,
+ "step": 568
+ },
+ {
+ "epoch": 14.34,
+ "learning_rate": 0.00027203525641025637,
+ "loss": 0.3544,
+ "step": 569
+ },
+ {
+ "epoch": 14.36,
+ "learning_rate": 0.00027163461538461537,
+ "loss": 0.3721,
+ "step": 570
+ },
+ {
+ "epoch": 14.39,
+ "learning_rate": 0.0002712339743589743,
+ "loss": 0.3769,
+ "step": 571
+ },
+ {
+ "epoch": 14.41,
+ "learning_rate": 0.0002708333333333333,
+ "loss": 0.3872,
+ "step": 572
+ },
+ {
+ "epoch": 14.44,
+ "learning_rate": 0.0002704326923076923,
+ "loss": 0.363,
+ "step": 573
+ },
+ {
+ "epoch": 14.46,
+ "learning_rate": 0.00027003205128205127,
+ "loss": 0.3842,
+ "step": 574
+ },
+ {
+ "epoch": 14.49,
+ "learning_rate": 0.00026963141025641027,
+ "loss": 0.3737,
+ "step": 575
+ },
+ {
+ "epoch": 14.51,
+ "learning_rate": 0.0002692307692307692,
+ "loss": 0.3981,
+ "step": 576
+ },
+ {
+ "epoch": 14.54,
+ "learning_rate": 0.0002688301282051282,
+ "loss": 0.3656,
+ "step": 577
+ },
+ {
+ "epoch": 14.56,
+ "learning_rate": 0.00026842948717948717,
+ "loss": 0.3852,
+ "step": 578
+ },
+ {
+ "epoch": 14.59,
+ "learning_rate": 0.00026802884615384617,
+ "loss": 0.3865,
+ "step": 579
+ },
+ {
+ "epoch": 14.61,
+ "learning_rate": 0.0002676282051282051,
+ "loss": 0.3747,
+ "step": 580
+ },
+ {
+ "epoch": 14.64,
+ "learning_rate": 0.0002672275641025641,
+ "loss": 0.3729,
+ "step": 581
+ },
+ {
+ "epoch": 14.66,
+ "learning_rate": 0.00026682692307692307,
+ "loss": 0.3785,
+ "step": 582
+ },
+ {
+ "epoch": 14.69,
+ "learning_rate": 0.00026642628205128207,
+ "loss": 0.4018,
+ "step": 583
+ },
+ {
+ "epoch": 14.71,
+ "learning_rate": 0.000266025641025641,
+ "loss": 0.3854,
+ "step": 584
+ },
+ {
+ "epoch": 14.74,
+ "learning_rate": 0.000265625,
+ "loss": 0.4161,
+ "step": 585
+ },
+ {
+ "epoch": 14.77,
+ "learning_rate": 0.00026522435897435897,
+ "loss": 0.4093,
+ "step": 586
+ },
+ {
+ "epoch": 14.79,
+ "learning_rate": 0.00026482371794871797,
+ "loss": 0.4126,
+ "step": 587
+ },
+ {
+ "epoch": 14.82,
+ "learning_rate": 0.0002644230769230769,
+ "loss": 0.3977,
+ "step": 588
+ },
+ {
+ "epoch": 14.84,
+ "learning_rate": 0.0002640224358974359,
+ "loss": 0.4286,
+ "step": 589
+ },
+ {
+ "epoch": 14.87,
+ "learning_rate": 0.00026362179487179487,
+ "loss": 0.4371,
+ "step": 590
+ },
+ {
+ "epoch": 14.89,
+ "learning_rate": 0.00026322115384615387,
+ "loss": 0.4066,
+ "step": 591
+ },
+ {
+ "epoch": 14.92,
+ "learning_rate": 0.0002628205128205128,
+ "loss": 0.433,
+ "step": 592
+ },
+ {
+ "epoch": 14.94,
+ "learning_rate": 0.0002624198717948718,
+ "loss": 0.4361,
+ "step": 593
+ },
+ {
+ "epoch": 14.97,
+ "learning_rate": 0.00026201923076923076,
+ "loss": 0.4276,
+ "step": 594
+ },
+ {
+ "epoch": 14.99,
+ "learning_rate": 0.00026161858974358977,
+ "loss": 0.4401,
+ "step": 595
+ },
+ {
+ "epoch": 15.02,
+ "learning_rate": 0.0002612179487179487,
+ "loss": 0.3514,
+ "step": 596
+ },
+ {
+ "epoch": 15.04,
+ "learning_rate": 0.0002608173076923077,
+ "loss": 0.306,
+ "step": 597
+ },
+ {
+ "epoch": 15.07,
+ "learning_rate": 0.0002604166666666667,
+ "loss": 0.2862,
+ "step": 598
+ },
+ {
+ "epoch": 15.09,
+ "learning_rate": 0.00026001602564102566,
+ "loss": 0.3042,
+ "step": 599
+ },
+ {
+ "epoch": 15.12,
+ "learning_rate": 0.00025961538461538467,
+ "loss": 0.2978,
+ "step": 600
+ },
+ {
+ "epoch": 15.14,
+ "learning_rate": 0.0002592147435897436,
+ "loss": 0.2975,
+ "step": 601
+ },
+ {
+ "epoch": 15.17,
+ "learning_rate": 0.0002588141025641026,
+ "loss": 0.3172,
+ "step": 602
+ },
+ {
+ "epoch": 15.19,
+ "learning_rate": 0.00025841346153846156,
+ "loss": 0.3046,
+ "step": 603
+ },
+ {
+ "epoch": 15.22,
+ "learning_rate": 0.00025801282051282057,
+ "loss": 0.3008,
+ "step": 604
+ },
+ {
+ "epoch": 15.24,
+ "learning_rate": 0.00025761217948717946,
+ "loss": 0.2956,
+ "step": 605
+ },
+ {
+ "epoch": 15.27,
+ "learning_rate": 0.00025721153846153846,
+ "loss": 0.3137,
+ "step": 606
+ },
+ {
+ "epoch": 15.29,
+ "learning_rate": 0.0002568108974358974,
+ "loss": 0.2961,
+ "step": 607
+ },
+ {
+ "epoch": 15.32,
+ "learning_rate": 0.0002564102564102564,
+ "loss": 0.3059,
+ "step": 608
+ },
+ {
+ "epoch": 15.34,
+ "learning_rate": 0.00025600961538461536,
+ "loss": 0.3077,
+ "step": 609
+ },
+ {
+ "epoch": 15.37,
+ "learning_rate": 0.00025560897435897436,
+ "loss": 0.3189,
+ "step": 610
+ },
+ {
+ "epoch": 15.4,
+ "learning_rate": 0.0002552083333333333,
+ "loss": 0.309,
+ "step": 611
+ },
+ {
+ "epoch": 15.42,
+ "learning_rate": 0.0002548076923076923,
+ "loss": 0.3324,
+ "step": 612
+ },
+ {
+ "epoch": 15.45,
+ "learning_rate": 0.00025440705128205126,
+ "loss": 0.3159,
+ "step": 613
+ },
+ {
+ "epoch": 15.47,
+ "learning_rate": 0.00025400641025641026,
+ "loss": 0.3188,
+ "step": 614
+ },
+ {
+ "epoch": 15.5,
+ "learning_rate": 0.0002536057692307692,
+ "loss": 0.3174,
+ "step": 615
+ },
+ {
+ "epoch": 15.52,
+ "learning_rate": 0.0002532051282051282,
+ "loss": 0.3074,
+ "step": 616
+ },
+ {
+ "epoch": 15.55,
+ "learning_rate": 0.00025280448717948716,
+ "loss": 0.3145,
+ "step": 617
+ },
+ {
+ "epoch": 15.57,
+ "learning_rate": 0.00025240384615384616,
+ "loss": 0.317,
+ "step": 618
+ },
+ {
+ "epoch": 15.6,
+ "learning_rate": 0.0002520032051282051,
+ "loss": 0.318,
+ "step": 619
+ },
+ {
+ "epoch": 15.62,
+ "learning_rate": 0.0002516025641025641,
+ "loss": 0.3236,
+ "step": 620
+ },
+ {
+ "epoch": 15.65,
+ "learning_rate": 0.00025120192307692305,
+ "loss": 0.3282,
+ "step": 621
+ },
+ {
+ "epoch": 15.67,
+ "learning_rate": 0.00025080128205128206,
+ "loss": 0.3394,
+ "step": 622
+ },
+ {
+ "epoch": 15.7,
+ "learning_rate": 0.000250400641025641,
+ "loss": 0.3278,
+ "step": 623
+ },
+ {
+ "epoch": 15.72,
+ "learning_rate": 0.00025,
+ "loss": 0.3255,
+ "step": 624
+ },
+ {
+ "epoch": 15.75,
+ "learning_rate": 0.000249599358974359,
+ "loss": 0.3256,
+ "step": 625
+ },
+ {
+ "epoch": 15.77,
+ "learning_rate": 0.00024919871794871795,
+ "loss": 0.3379,
+ "step": 626
+ },
+ {
+ "epoch": 15.8,
+ "learning_rate": 0.00024879807692307696,
+ "loss": 0.3384,
+ "step": 627
+ },
+ {
+ "epoch": 15.82,
+ "learning_rate": 0.0002483974358974359,
+ "loss": 0.3483,
+ "step": 628
+ },
+ {
+ "epoch": 15.85,
+ "learning_rate": 0.0002479967948717949,
+ "loss": 0.3478,
+ "step": 629
+ },
+ {
+ "epoch": 15.87,
+ "learning_rate": 0.00024759615384615385,
+ "loss": 0.337,
+ "step": 630
+ },
+ {
+ "epoch": 15.9,
+ "learning_rate": 0.00024719551282051286,
+ "loss": 0.3287,
+ "step": 631
+ },
+ {
+ "epoch": 15.92,
+ "learning_rate": 0.0002467948717948718,
+ "loss": 0.3478,
+ "step": 632
+ },
+ {
+ "epoch": 15.95,
+ "learning_rate": 0.0002463942307692308,
+ "loss": 0.3567,
+ "step": 633
+ },
+ {
+ "epoch": 15.97,
+ "learning_rate": 0.00024599358974358975,
+ "loss": 0.3359,
+ "step": 634
+ },
+ {
+ "epoch": 16.0,
+ "learning_rate": 0.0002455929487179487,
+ "loss": 0.3492,
+ "step": 635
+ },
+ {
+ "epoch": 16.03,
+ "learning_rate": 0.0002451923076923077,
+ "loss": 0.2497,
+ "step": 636
+ },
+ {
+ "epoch": 16.05,
+ "learning_rate": 0.00024479166666666665,
+ "loss": 0.2416,
+ "step": 637
+ },
+ {
+ "epoch": 16.08,
+ "learning_rate": 0.00024439102564102565,
+ "loss": 0.2576,
+ "step": 638
+ },
+ {
+ "epoch": 16.1,
+ "learning_rate": 0.00024399038461538463,
+ "loss": 0.243,
+ "step": 639
+ },
+ {
+ "epoch": 16.13,
+ "learning_rate": 0.0002435897435897436,
+ "loss": 0.2527,
+ "step": 640
+ },
+ {
+ "epoch": 16.15,
+ "learning_rate": 0.00024318910256410258,
+ "loss": 0.2606,
+ "step": 641
+ },
+ {
+ "epoch": 16.18,
+ "learning_rate": 0.00024278846153846155,
+ "loss": 0.2555,
+ "step": 642
+ },
+ {
+ "epoch": 16.2,
+ "learning_rate": 0.00024238782051282052,
+ "loss": 0.2581,
+ "step": 643
+ },
+ {
+ "epoch": 16.23,
+ "learning_rate": 0.0002419871794871795,
+ "loss": 0.2508,
+ "step": 644
+ },
+ {
+ "epoch": 16.25,
+ "learning_rate": 0.00024158653846153847,
+ "loss": 0.281,
+ "step": 645
+ },
+ {
+ "epoch": 16.28,
+ "learning_rate": 0.00024118589743589745,
+ "loss": 0.2626,
+ "step": 646
+ },
+ {
+ "epoch": 16.3,
+ "learning_rate": 0.00024078525641025642,
+ "loss": 0.2668,
+ "step": 647
+ },
+ {
+ "epoch": 16.33,
+ "learning_rate": 0.0002403846153846154,
+ "loss": 0.2675,
+ "step": 648
+ },
+ {
+ "epoch": 16.35,
+ "learning_rate": 0.00023998397435897437,
+ "loss": 0.2491,
+ "step": 649
+ },
+ {
+ "epoch": 16.38,
+ "learning_rate": 0.00023958333333333335,
+ "loss": 0.2459,
+ "step": 650
+ },
+ {
+ "epoch": 16.4,
+ "learning_rate": 0.00023918269230769232,
+ "loss": 0.2515,
+ "step": 651
+ },
+ {
+ "epoch": 16.43,
+ "learning_rate": 0.0002387820512820513,
+ "loss": 0.2612,
+ "step": 652
+ },
+ {
+ "epoch": 16.45,
+ "learning_rate": 0.00023838141025641027,
+ "loss": 0.2612,
+ "step": 653
+ },
+ {
+ "epoch": 16.48,
+ "learning_rate": 0.00023798076923076922,
+ "loss": 0.265,
+ "step": 654
+ },
+ {
+ "epoch": 16.5,
+ "learning_rate": 0.0002375801282051282,
+ "loss": 0.2634,
+ "step": 655
+ },
+ {
+ "epoch": 16.53,
+ "learning_rate": 0.00023717948717948717,
+ "loss": 0.2839,
+ "step": 656
+ },
+ {
+ "epoch": 16.55,
+ "learning_rate": 0.00023677884615384614,
+ "loss": 0.2595,
+ "step": 657
+ },
+ {
+ "epoch": 16.58,
+ "learning_rate": 0.00023637820512820512,
+ "loss": 0.2764,
+ "step": 658
+ },
+ {
+ "epoch": 16.6,
+ "learning_rate": 0.0002359775641025641,
+ "loss": 0.2553,
+ "step": 659
+ },
+ {
+ "epoch": 16.63,
+ "learning_rate": 0.00023557692307692307,
+ "loss": 0.2716,
+ "step": 660
+ },
+ {
+ "epoch": 16.66,
+ "learning_rate": 0.00023517628205128204,
+ "loss": 0.289,
+ "step": 661
+ },
+ {
+ "epoch": 16.68,
+ "learning_rate": 0.00023477564102564102,
+ "loss": 0.2483,
+ "step": 662
+ },
+ {
+ "epoch": 16.71,
+ "learning_rate": 0.000234375,
+ "loss": 0.269,
+ "step": 663
+ },
+ {
+ "epoch": 16.73,
+ "learning_rate": 0.000233974358974359,
+ "loss": 0.2738,
+ "step": 664
+ },
+ {
+ "epoch": 16.76,
+ "learning_rate": 0.00023357371794871797,
+ "loss": 0.2646,
+ "step": 665
+ },
+ {
+ "epoch": 16.78,
+ "learning_rate": 0.00023317307692307694,
+ "loss": 0.2793,
+ "step": 666
+ },
+ {
+ "epoch": 16.81,
+ "learning_rate": 0.00023277243589743592,
+ "loss": 0.2666,
+ "step": 667
+ },
+ {
+ "epoch": 16.83,
+ "learning_rate": 0.0002323717948717949,
+ "loss": 0.2799,
+ "step": 668
+ },
+ {
+ "epoch": 16.86,
+ "learning_rate": 0.00023197115384615387,
+ "loss": 0.2677,
+ "step": 669
+ },
+ {
+ "epoch": 16.88,
+ "learning_rate": 0.00023157051282051284,
+ "loss": 0.2889,
+ "step": 670
+ },
+ {
+ "epoch": 16.91,
+ "learning_rate": 0.00023116987179487182,
+ "loss": 0.2869,
+ "step": 671
+ },
+ {
+ "epoch": 16.93,
+ "learning_rate": 0.0002307692307692308,
+ "loss": 0.2814,
+ "step": 672
+ },
+ {
+ "epoch": 16.96,
+ "learning_rate": 0.00023036858974358974,
+ "loss": 0.2881,
+ "step": 673
+ },
+ {
+ "epoch": 16.98,
+ "learning_rate": 0.0002299679487179487,
+ "loss": 0.2777,
+ "step": 674
+ },
+ {
+ "epoch": 17.01,
+ "learning_rate": 0.0002295673076923077,
+ "loss": 0.2713,
+ "step": 675
+ },
+ {
+ "epoch": 17.03,
+ "learning_rate": 0.00022916666666666666,
+ "loss": 0.2176,
+ "step": 676
+ },
+ {
+ "epoch": 17.06,
+ "learning_rate": 0.00022876602564102564,
+ "loss": 0.2032,
+ "step": 677
+ },
+ {
+ "epoch": 17.08,
+ "learning_rate": 0.0002283653846153846,
+ "loss": 0.2207,
+ "step": 678
+ },
+ {
+ "epoch": 17.11,
+ "learning_rate": 0.0002279647435897436,
+ "loss": 0.2183,
+ "step": 679
+ },
+ {
+ "epoch": 17.13,
+ "learning_rate": 0.00022756410256410256,
+ "loss": 0.2238,
+ "step": 680
+ },
+ {
+ "epoch": 17.16,
+ "learning_rate": 0.00022716346153846154,
+ "loss": 0.2126,
+ "step": 681
+ },
+ {
+ "epoch": 17.18,
+ "learning_rate": 0.0002267628205128205,
+ "loss": 0.2104,
+ "step": 682
+ },
+ {
+ "epoch": 17.21,
+ "learning_rate": 0.00022636217948717949,
+ "loss": 0.2151,
+ "step": 683
+ },
+ {
+ "epoch": 17.23,
+ "learning_rate": 0.00022596153846153846,
+ "loss": 0.2093,
+ "step": 684
+ },
+ {
+ "epoch": 17.26,
+ "learning_rate": 0.00022556089743589743,
+ "loss": 0.2308,
+ "step": 685
+ },
+ {
+ "epoch": 17.29,
+ "learning_rate": 0.0002251602564102564,
+ "loss": 0.2276,
+ "step": 686
+ },
+ {
+ "epoch": 17.31,
+ "learning_rate": 0.00022475961538461538,
+ "loss": 0.2328,
+ "step": 687
+ },
+ {
+ "epoch": 17.34,
+ "learning_rate": 0.00022435897435897436,
+ "loss": 0.2076,
+ "step": 688
+ },
+ {
+ "epoch": 17.36,
+ "learning_rate": 0.00022395833333333336,
+ "loss": 0.2125,
+ "step": 689
+ },
+ {
+ "epoch": 17.39,
+ "learning_rate": 0.00022355769230769234,
+ "loss": 0.2181,
+ "step": 690
+ },
+ {
+ "epoch": 17.41,
+ "learning_rate": 0.0002231570512820513,
+ "loss": 0.2153,
+ "step": 691
+ },
+ {
+ "epoch": 17.44,
+ "learning_rate": 0.00022275641025641028,
+ "loss": 0.2122,
+ "step": 692
+ },
+ {
+ "epoch": 17.46,
+ "learning_rate": 0.00022235576923076923,
+ "loss": 0.2262,
+ "step": 693
+ },
+ {
+ "epoch": 17.49,
+ "learning_rate": 0.0002219551282051282,
+ "loss": 0.2206,
+ "step": 694
+ },
+ {
+ "epoch": 17.51,
+ "learning_rate": 0.00022155448717948718,
+ "loss": 0.2223,
+ "step": 695
+ },
+ {
+ "epoch": 17.54,
+ "learning_rate": 0.00022115384615384616,
+ "loss": 0.2306,
+ "step": 696
+ },
+ {
+ "epoch": 17.56,
+ "learning_rate": 0.00022075320512820513,
+ "loss": 0.213,
+ "step": 697
+ },
+ {
+ "epoch": 17.59,
+ "learning_rate": 0.0002203525641025641,
+ "loss": 0.2198,
+ "step": 698
+ },
+ {
+ "epoch": 17.61,
+ "learning_rate": 0.00021995192307692308,
+ "loss": 0.2299,
+ "step": 699
+ },
+ {
+ "epoch": 17.64,
+ "learning_rate": 0.00021955128205128206,
+ "loss": 0.2213,
+ "step": 700
+ },
+ {
+ "epoch": 17.66,
+ "learning_rate": 0.00021915064102564103,
+ "loss": 0.2183,
+ "step": 701
+ },
+ {
+ "epoch": 17.69,
+ "learning_rate": 0.00021875,
+ "loss": 0.2301,
+ "step": 702
+ },
+ {
+ "epoch": 17.71,
+ "learning_rate": 0.00021834935897435898,
+ "loss": 0.2348,
+ "step": 703
+ },
+ {
+ "epoch": 17.74,
+ "learning_rate": 0.00021794871794871795,
+ "loss": 0.229,
+ "step": 704
+ },
+ {
+ "epoch": 17.76,
+ "learning_rate": 0.00021754807692307693,
+ "loss": 0.2397,
+ "step": 705
+ },
+ {
+ "epoch": 17.79,
+ "learning_rate": 0.0002171474358974359,
+ "loss": 0.2254,
+ "step": 706
+ },
+ {
+ "epoch": 17.81,
+ "learning_rate": 0.00021674679487179488,
+ "loss": 0.2306,
+ "step": 707
+ },
+ {
+ "epoch": 17.84,
+ "learning_rate": 0.00021634615384615385,
+ "loss": 0.2388,
+ "step": 708
+ },
+ {
+ "epoch": 17.86,
+ "learning_rate": 0.00021594551282051283,
+ "loss": 0.2374,
+ "step": 709
+ },
+ {
+ "epoch": 17.89,
+ "learning_rate": 0.0002155448717948718,
+ "loss": 0.2412,
+ "step": 710
+ },
+ {
+ "epoch": 17.91,
+ "learning_rate": 0.00021514423076923078,
+ "loss": 0.2203,
+ "step": 711
+ },
+ {
+ "epoch": 17.94,
+ "learning_rate": 0.00021474358974358972,
+ "loss": 0.2359,
+ "step": 712
+ },
+ {
+ "epoch": 17.97,
+ "learning_rate": 0.0002143429487179487,
+ "loss": 0.2343,
+ "step": 713
+ },
+ {
+ "epoch": 17.99,
+ "learning_rate": 0.00021394230769230767,
+ "loss": 0.2397,
+ "step": 714
+ },
+ {
+ "epoch": 18.02,
+ "learning_rate": 0.00021354166666666665,
+ "loss": 0.2055,
+ "step": 715
+ },
+ {
+ "epoch": 18.04,
+ "learning_rate": 0.00021314102564102565,
+ "loss": 0.1723,
+ "step": 716
+ },
+ {
+ "epoch": 18.07,
+ "learning_rate": 0.00021274038461538463,
+ "loss": 0.1782,
+ "step": 717
+ },
+ {
+ "epoch": 18.09,
+ "learning_rate": 0.0002123397435897436,
+ "loss": 0.1813,
+ "step": 718
+ },
+ {
+ "epoch": 18.12,
+ "learning_rate": 0.00021193910256410257,
+ "loss": 0.1834,
+ "step": 719
+ },
+ {
+ "epoch": 18.14,
+ "learning_rate": 0.00021153846153846155,
+ "loss": 0.1759,
+ "step": 720
+ },
+ {
+ "epoch": 18.17,
+ "learning_rate": 0.00021113782051282052,
+ "loss": 0.1855,
+ "step": 721
+ },
+ {
+ "epoch": 18.19,
+ "learning_rate": 0.0002107371794871795,
+ "loss": 0.1756,
+ "step": 722
+ },
+ {
+ "epoch": 18.22,
+ "learning_rate": 0.00021033653846153847,
+ "loss": 0.1783,
+ "step": 723
+ },
+ {
+ "epoch": 18.24,
+ "learning_rate": 0.00020993589743589745,
+ "loss": 0.1899,
+ "step": 724
+ },
+ {
+ "epoch": 18.27,
+ "learning_rate": 0.00020953525641025642,
+ "loss": 0.1817,
+ "step": 725
+ },
+ {
+ "epoch": 18.29,
+ "learning_rate": 0.0002091346153846154,
+ "loss": 0.1856,
+ "step": 726
+ },
+ {
+ "epoch": 18.32,
+ "learning_rate": 0.00020873397435897437,
+ "loss": 0.1849,
+ "step": 727
+ },
+ {
+ "epoch": 18.34,
+ "learning_rate": 0.00020833333333333335,
+ "loss": 0.1758,
+ "step": 728
+ },
+ {
+ "epoch": 18.37,
+ "learning_rate": 0.00020793269230769232,
+ "loss": 0.1907,
+ "step": 729
+ },
+ {
+ "epoch": 18.39,
+ "learning_rate": 0.0002075320512820513,
+ "loss": 0.1836,
+ "step": 730
+ },
+ {
+ "epoch": 18.42,
+ "learning_rate": 0.00020713141025641027,
+ "loss": 0.1828,
+ "step": 731
+ },
+ {
+ "epoch": 18.44,
+ "learning_rate": 0.00020673076923076922,
+ "loss": 0.1916,
+ "step": 732
+ },
+ {
+ "epoch": 18.47,
+ "learning_rate": 0.0002063301282051282,
+ "loss": 0.1849,
+ "step": 733
+ },
+ {
+ "epoch": 18.49,
+ "learning_rate": 0.00020592948717948717,
+ "loss": 0.1929,
+ "step": 734
+ },
+ {
+ "epoch": 18.52,
+ "learning_rate": 0.00020552884615384614,
+ "loss": 0.1779,
+ "step": 735
+ },
+ {
+ "epoch": 18.54,
+ "learning_rate": 0.00020512820512820512,
+ "loss": 0.1805,
+ "step": 736
+ },
+ {
+ "epoch": 18.57,
+ "learning_rate": 0.0002047275641025641,
+ "loss": 0.1891,
+ "step": 737
+ },
+ {
+ "epoch": 18.6,
+ "learning_rate": 0.00020432692307692307,
+ "loss": 0.1932,
+ "step": 738
+ },
+ {
+ "epoch": 18.62,
+ "learning_rate": 0.00020392628205128204,
+ "loss": 0.1923,
+ "step": 739
+ },
+ {
+ "epoch": 18.65,
+ "learning_rate": 0.00020352564102564102,
+ "loss": 0.1882,
+ "step": 740
+ },
+ {
+ "epoch": 18.67,
+ "learning_rate": 0.00020312500000000002,
+ "loss": 0.2012,
+ "step": 741
+ },
+ {
+ "epoch": 18.7,
+ "learning_rate": 0.000202724358974359,
+ "loss": 0.1963,
+ "step": 742
+ },
+ {
+ "epoch": 18.72,
+ "learning_rate": 0.00020232371794871797,
+ "loss": 0.202,
+ "step": 743
+ },
+ {
+ "epoch": 18.75,
+ "learning_rate": 0.00020192307692307694,
+ "loss": 0.1996,
+ "step": 744
+ },
+ {
+ "epoch": 18.77,
+ "learning_rate": 0.00020152243589743592,
+ "loss": 0.2056,
+ "step": 745
+ },
+ {
+ "epoch": 18.8,
+ "learning_rate": 0.0002011217948717949,
+ "loss": 0.1946,
+ "step": 746
+ },
+ {
+ "epoch": 18.82,
+ "learning_rate": 0.00020072115384615387,
+ "loss": 0.1954,
+ "step": 747
+ },
+ {
+ "epoch": 18.85,
+ "learning_rate": 0.00020032051282051284,
+ "loss": 0.1966,
+ "step": 748
+ },
+ {
+ "epoch": 18.87,
+ "learning_rate": 0.00019991987179487182,
+ "loss": 0.2077,
+ "step": 749
+ },
+ {
+ "epoch": 18.9,
+ "learning_rate": 0.0001995192307692308,
+ "loss": 0.2,
+ "step": 750
+ },
+ {
+ "epoch": 18.92,
+ "learning_rate": 0.00019911858974358974,
+ "loss": 0.2092,
+ "step": 751
+ },
+ {
+ "epoch": 18.95,
+ "learning_rate": 0.0001987179487179487,
+ "loss": 0.2023,
+ "step": 752
+ },
+ {
+ "epoch": 18.97,
+ "learning_rate": 0.0001983173076923077,
+ "loss": 0.2056,
+ "step": 753
+ },
+ {
+ "epoch": 19.0,
+ "learning_rate": 0.00019791666666666666,
+ "loss": 0.2059,
+ "step": 754
+ },
+ {
+ "epoch": 19.02,
+ "learning_rate": 0.00019751602564102564,
+ "loss": 0.1581,
+ "step": 755
+ },
+ {
+ "epoch": 19.05,
+ "learning_rate": 0.0001971153846153846,
+ "loss": 0.155,
+ "step": 756
+ },
+ {
+ "epoch": 19.07,
+ "learning_rate": 0.00019671474358974359,
+ "loss": 0.156,
+ "step": 757
+ },
+ {
+ "epoch": 19.1,
+ "learning_rate": 0.00019631410256410256,
+ "loss": 0.161,
+ "step": 758
+ },
+ {
+ "epoch": 19.12,
+ "learning_rate": 0.00019591346153846154,
+ "loss": 0.1608,
+ "step": 759
+ },
+ {
+ "epoch": 19.15,
+ "learning_rate": 0.0001955128205128205,
+ "loss": 0.1576,
+ "step": 760
+ },
+ {
+ "epoch": 19.17,
+ "learning_rate": 0.00019511217948717948,
+ "loss": 0.1539,
+ "step": 761
+ },
+ {
+ "epoch": 19.2,
+ "learning_rate": 0.00019471153846153846,
+ "loss": 0.1549,
+ "step": 762
+ },
+ {
+ "epoch": 19.23,
+ "learning_rate": 0.00019431089743589743,
+ "loss": 0.1593,
+ "step": 763
+ },
+ {
+ "epoch": 19.25,
+ "learning_rate": 0.0001939102564102564,
+ "loss": 0.162,
+ "step": 764
+ },
+ {
+ "epoch": 19.28,
+ "learning_rate": 0.00019350961538461538,
+ "loss": 0.1591,
+ "step": 765
+ },
+ {
+ "epoch": 19.3,
+ "learning_rate": 0.00019310897435897436,
+ "loss": 0.1569,
+ "step": 766
+ },
+ {
+ "epoch": 19.33,
+ "learning_rate": 0.00019270833333333336,
+ "loss": 0.1628,
+ "step": 767
+ },
+ {
+ "epoch": 19.35,
+ "learning_rate": 0.00019230769230769233,
+ "loss": 0.1529,
+ "step": 768
+ },
+ {
+ "epoch": 19.38,
+ "learning_rate": 0.0001919070512820513,
+ "loss": 0.1606,
+ "step": 769
+ },
+ {
+ "epoch": 19.4,
+ "learning_rate": 0.00019150641025641028,
+ "loss": 0.1564,
+ "step": 770
+ },
+ {
+ "epoch": 19.43,
+ "learning_rate": 0.00019110576923076923,
+ "loss": 0.1599,
+ "step": 771
+ },
+ {
+ "epoch": 19.45,
+ "learning_rate": 0.0001907051282051282,
+ "loss": 0.1653,
+ "step": 772
+ },
+ {
+ "epoch": 19.48,
+ "learning_rate": 0.00019030448717948718,
+ "loss": 0.1587,
+ "step": 773
+ },
+ {
+ "epoch": 19.5,
+ "learning_rate": 0.00018990384615384616,
+ "loss": 0.1683,
+ "step": 774
+ },
+ {
+ "epoch": 19.53,
+ "learning_rate": 0.00018950320512820513,
+ "loss": 0.157,
+ "step": 775
+ },
+ {
+ "epoch": 19.55,
+ "learning_rate": 0.0001891025641025641,
+ "loss": 0.1588,
+ "step": 776
+ },
+ {
+ "epoch": 19.58,
+ "learning_rate": 0.00018870192307692308,
+ "loss": 0.1709,
+ "step": 777
+ },
+ {
+ "epoch": 19.6,
+ "learning_rate": 0.00018830128205128205,
+ "loss": 0.1635,
+ "step": 778
+ },
+ {
+ "epoch": 19.63,
+ "learning_rate": 0.00018790064102564103,
+ "loss": 0.1581,
+ "step": 779
+ },
+ {
+ "epoch": 19.65,
+ "learning_rate": 0.0001875,
+ "loss": 0.1664,
+ "step": 780
+ },
+ {
+ "epoch": 19.68,
+ "learning_rate": 0.00018709935897435898,
+ "loss": 0.1712,
+ "step": 781
+ },
+ {
+ "epoch": 19.7,
+ "learning_rate": 0.00018669871794871795,
+ "loss": 0.165,
+ "step": 782
+ },
+ {
+ "epoch": 19.73,
+ "learning_rate": 0.00018629807692307693,
+ "loss": 0.1692,
+ "step": 783
+ },
+ {
+ "epoch": 19.75,
+ "learning_rate": 0.0001858974358974359,
+ "loss": 0.1658,
+ "step": 784
+ },
+ {
+ "epoch": 19.78,
+ "learning_rate": 0.00018549679487179488,
+ "loss": 0.1702,
+ "step": 785
+ },
+ {
+ "epoch": 19.8,
+ "learning_rate": 0.00018509615384615385,
+ "loss": 0.1672,
+ "step": 786
+ },
+ {
+ "epoch": 19.83,
+ "learning_rate": 0.00018469551282051283,
+ "loss": 0.1708,
+ "step": 787
+ },
+ {
+ "epoch": 19.86,
+ "learning_rate": 0.0001842948717948718,
+ "loss": 0.1705,
+ "step": 788
+ },
+ {
+ "epoch": 19.88,
+ "learning_rate": 0.00018389423076923078,
+ "loss": 0.1853,
+ "step": 789
+ },
+ {
+ "epoch": 19.91,
+ "learning_rate": 0.00018349358974358972,
+ "loss": 0.1701,
+ "step": 790
+ },
+ {
+ "epoch": 19.93,
+ "learning_rate": 0.0001830929487179487,
+ "loss": 0.1726,
+ "step": 791
+ },
+ {
+ "epoch": 19.96,
+ "learning_rate": 0.00018269230769230767,
+ "loss": 0.1829,
+ "step": 792
+ },
+ {
+ "epoch": 19.98,
+ "learning_rate": 0.00018229166666666665,
+ "loss": 0.1749,
+ "step": 793
+ },
+ {
+ "epoch": 20.01,
+ "learning_rate": 0.00018189102564102565,
+ "loss": 0.1559,
+ "step": 794
+ },
+ {
+ "epoch": 20.03,
+ "learning_rate": 0.00018149038461538462,
+ "loss": 0.1424,
+ "step": 795
+ },
+ {
+ "epoch": 20.06,
+ "learning_rate": 0.0001810897435897436,
+ "loss": 0.1386,
+ "step": 796
+ },
+ {
+ "epoch": 20.08,
+ "learning_rate": 0.00018068910256410257,
+ "loss": 0.1388,
+ "step": 797
+ },
+ {
+ "epoch": 20.11,
+ "learning_rate": 0.00018028846153846155,
+ "loss": 0.1385,
+ "step": 798
+ },
+ {
+ "epoch": 20.13,
+ "learning_rate": 0.00017988782051282052,
+ "loss": 0.1349,
+ "step": 799
+ },
+ {
+ "epoch": 20.16,
+ "learning_rate": 0.0001794871794871795,
+ "loss": 0.1357,
+ "step": 800
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 4.5029205856567296e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-800/training_args.bin b/checkpoint-800/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-800/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/checkpoint-900/README.md b/checkpoint-900/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..08371015f02382e6fcba318f4aaea54ae52cd3c4
--- /dev/null
+++ b/checkpoint-900/README.md
@@ -0,0 +1,34 @@
+---
+library_name: peft
+---
+## Training procedure
+
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+
+The following `bitsandbytes` quantization config was used during training:
+- quant_method: bitsandbytes
+- load_in_8bit: True
+- load_in_4bit: False
+- llm_int8_threshold: 6.0
+- llm_int8_skip_modules: None
+- llm_int8_enable_fp32_cpu_offload: False
+- llm_int8_has_fp16_weight: False
+- bnb_4bit_quant_type: fp4
+- bnb_4bit_use_double_quant: False
+- bnb_4bit_compute_dtype: float32
+### Framework versions
+
+- PEFT 0.6.0.dev0
+
+- PEFT 0.6.0.dev0
diff --git a/checkpoint-900/adapter_config.json b/checkpoint-900/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..751d838ac0c1ae5ca71ca448b25d7a8a0173f01b
--- /dev/null
+++ b/checkpoint-900/adapter_config.json
@@ -0,0 +1,23 @@
+{
+ "auto_mapping": null,
+ "base_model_name_or_path": "bigscience/bloomz-3b",
+ "bias": "none",
+ "fan_in_fan_out": false,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "lora_alpha": 16,
+ "lora_dropout": 0.0,
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 8,
+ "revision": null,
+ "target_modules": [
+ "dense_4h_to_h",
+ "dense",
+ "dense_h_to_4h",
+ "query_key_value"
+ ],
+ "task_type": "CAUSAL_LM"
+}
\ No newline at end of file
diff --git a/checkpoint-900/adapter_model.bin b/checkpoint-900/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..01e8ee26735b808ebecd882fddf3f7947200e251
--- /dev/null
+++ b/checkpoint-900/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ad64c4d8d47a306b5cd84cdfbd03afce3f5f437925dd6eda4dc9d73545b47603
+size 39409357
diff --git a/checkpoint-900/optimizer.pt b/checkpoint-900/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..2ce29d104db33cda8e8669aba93c5c9448b23f4b
--- /dev/null
+++ b/checkpoint-900/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:957c95e1d45a7702b543e1041f9fef2e51e692517e291bafb828e8c107d84682
+size 78844421
diff --git a/checkpoint-900/rng_state.pth b/checkpoint-900/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..6c28f5beeeb3e640e81e05216ffbd6b8723d5723
--- /dev/null
+++ b/checkpoint-900/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8a2ce2e2326ca04fc964861061399d5335411f688384a0f62f8f943f1fff584e
+size 14575
diff --git a/checkpoint-900/scheduler.pt b/checkpoint-900/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..7b112c5a6eed6444d140b8ad3a7ca1c3db70c16d
--- /dev/null
+++ b/checkpoint-900/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0cc075a74e1aaa5c87a6a5b75fea9b7c0b9ef8a99c4e5ed04bc0069b3c22636b
+size 627
diff --git a/checkpoint-900/special_tokens_map.json b/checkpoint-900/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..fdafe480f024ff444c7492147536765ce5d55a2d
--- /dev/null
+++ b/checkpoint-900/special_tokens_map.json
@@ -0,0 +1,6 @@
+{
+ "bos_token": "",
+ "eos_token": "",
+ "pad_token": "",
+ "unk_token": ""
+}
diff --git a/checkpoint-900/tokenizer.json b/checkpoint-900/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..dbf002cafbd4818dcff2abc9156c088d681b4533
--- /dev/null
+++ b/checkpoint-900/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
+size 14500471
diff --git a/checkpoint-900/tokenizer_config.json b/checkpoint-900/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b56cc9c2965c07132c35df3e2972e93d98c82c3
--- /dev/null
+++ b/checkpoint-900/tokenizer_config.json
@@ -0,0 +1,10 @@
+{
+ "add_prefix_space": false,
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "tokenizer_class": "BloomTokenizer",
+ "unk_token": ""
+}
diff --git a/checkpoint-900/trainer_state.json b/checkpoint-900/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..0654fc8dbd320641391e5153de81666b26138f8c
--- /dev/null
+++ b/checkpoint-900/trainer_state.json
@@ -0,0 +1,5419 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 22.677165354330707,
+ "eval_steps": 500,
+ "global_step": 900,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "learning_rate": 0.000499599358974359,
+ "loss": 3.1389,
+ "step": 1
+ },
+ {
+ "epoch": 0.05,
+ "learning_rate": 0.000499198717948718,
+ "loss": 2.8569,
+ "step": 2
+ },
+ {
+ "epoch": 0.08,
+ "learning_rate": 0.0004987980769230769,
+ "loss": 2.75,
+ "step": 3
+ },
+ {
+ "epoch": 0.1,
+ "learning_rate": 0.0004983974358974359,
+ "loss": 2.7375,
+ "step": 4
+ },
+ {
+ "epoch": 0.13,
+ "learning_rate": 0.0004979967948717949,
+ "loss": 2.7342,
+ "step": 5
+ },
+ {
+ "epoch": 0.15,
+ "learning_rate": 0.0004975961538461539,
+ "loss": 2.753,
+ "step": 6
+ },
+ {
+ "epoch": 0.18,
+ "learning_rate": 0.0004971955128205128,
+ "loss": 2.7124,
+ "step": 7
+ },
+ {
+ "epoch": 0.2,
+ "learning_rate": 0.0004967948717948718,
+ "loss": 2.7038,
+ "step": 8
+ },
+ {
+ "epoch": 0.23,
+ "learning_rate": 0.0004963942307692308,
+ "loss": 2.6142,
+ "step": 9
+ },
+ {
+ "epoch": 0.25,
+ "learning_rate": 0.0004959935897435898,
+ "loss": 2.6427,
+ "step": 10
+ },
+ {
+ "epoch": 0.28,
+ "learning_rate": 0.0004955929487179487,
+ "loss": 2.6265,
+ "step": 11
+ },
+ {
+ "epoch": 0.3,
+ "learning_rate": 0.0004951923076923077,
+ "loss": 2.6352,
+ "step": 12
+ },
+ {
+ "epoch": 0.33,
+ "learning_rate": 0.0004947916666666667,
+ "loss": 2.616,
+ "step": 13
+ },
+ {
+ "epoch": 0.35,
+ "learning_rate": 0.0004943910256410257,
+ "loss": 2.6937,
+ "step": 14
+ },
+ {
+ "epoch": 0.38,
+ "learning_rate": 0.0004939903846153846,
+ "loss": 2.6719,
+ "step": 15
+ },
+ {
+ "epoch": 0.4,
+ "learning_rate": 0.0004935897435897436,
+ "loss": 2.6043,
+ "step": 16
+ },
+ {
+ "epoch": 0.43,
+ "learning_rate": 0.0004931891025641026,
+ "loss": 2.565,
+ "step": 17
+ },
+ {
+ "epoch": 0.45,
+ "learning_rate": 0.0004927884615384616,
+ "loss": 2.5776,
+ "step": 18
+ },
+ {
+ "epoch": 0.48,
+ "learning_rate": 0.0004923878205128205,
+ "loss": 2.6326,
+ "step": 19
+ },
+ {
+ "epoch": 0.5,
+ "learning_rate": 0.0004919871794871795,
+ "loss": 2.5868,
+ "step": 20
+ },
+ {
+ "epoch": 0.53,
+ "learning_rate": 0.0004915865384615384,
+ "loss": 2.5499,
+ "step": 21
+ },
+ {
+ "epoch": 0.55,
+ "learning_rate": 0.0004911858974358974,
+ "loss": 2.5823,
+ "step": 22
+ },
+ {
+ "epoch": 0.58,
+ "learning_rate": 0.0004907852564102564,
+ "loss": 2.5858,
+ "step": 23
+ },
+ {
+ "epoch": 0.6,
+ "learning_rate": 0.0004903846153846154,
+ "loss": 2.5688,
+ "step": 24
+ },
+ {
+ "epoch": 0.63,
+ "learning_rate": 0.0004899839743589743,
+ "loss": 2.57,
+ "step": 25
+ },
+ {
+ "epoch": 0.66,
+ "learning_rate": 0.0004895833333333333,
+ "loss": 2.5569,
+ "step": 26
+ },
+ {
+ "epoch": 0.68,
+ "learning_rate": 0.0004891826923076923,
+ "loss": 2.5067,
+ "step": 27
+ },
+ {
+ "epoch": 0.71,
+ "learning_rate": 0.0004887820512820513,
+ "loss": 2.4936,
+ "step": 28
+ },
+ {
+ "epoch": 0.73,
+ "learning_rate": 0.0004883814102564103,
+ "loss": 2.5958,
+ "step": 29
+ },
+ {
+ "epoch": 0.76,
+ "learning_rate": 0.00048798076923076925,
+ "loss": 2.5213,
+ "step": 30
+ },
+ {
+ "epoch": 0.78,
+ "learning_rate": 0.0004875801282051282,
+ "loss": 2.6369,
+ "step": 31
+ },
+ {
+ "epoch": 0.81,
+ "learning_rate": 0.0004871794871794872,
+ "loss": 2.578,
+ "step": 32
+ },
+ {
+ "epoch": 0.83,
+ "learning_rate": 0.00048677884615384615,
+ "loss": 2.6214,
+ "step": 33
+ },
+ {
+ "epoch": 0.86,
+ "learning_rate": 0.00048637820512820515,
+ "loss": 2.6193,
+ "step": 34
+ },
+ {
+ "epoch": 0.88,
+ "learning_rate": 0.0004859775641025641,
+ "loss": 2.5948,
+ "step": 35
+ },
+ {
+ "epoch": 0.91,
+ "learning_rate": 0.0004855769230769231,
+ "loss": 2.5615,
+ "step": 36
+ },
+ {
+ "epoch": 0.93,
+ "learning_rate": 0.00048517628205128205,
+ "loss": 2.5246,
+ "step": 37
+ },
+ {
+ "epoch": 0.96,
+ "learning_rate": 0.00048477564102564105,
+ "loss": 2.4585,
+ "step": 38
+ },
+ {
+ "epoch": 0.98,
+ "learning_rate": 0.000484375,
+ "loss": 2.5621,
+ "step": 39
+ },
+ {
+ "epoch": 1.01,
+ "learning_rate": 0.000483974358974359,
+ "loss": 2.571,
+ "step": 40
+ },
+ {
+ "epoch": 1.03,
+ "learning_rate": 0.00048357371794871795,
+ "loss": 2.5382,
+ "step": 41
+ },
+ {
+ "epoch": 1.06,
+ "learning_rate": 0.00048317307692307695,
+ "loss": 2.5028,
+ "step": 42
+ },
+ {
+ "epoch": 1.08,
+ "learning_rate": 0.0004827724358974359,
+ "loss": 2.4812,
+ "step": 43
+ },
+ {
+ "epoch": 1.11,
+ "learning_rate": 0.0004823717948717949,
+ "loss": 2.5161,
+ "step": 44
+ },
+ {
+ "epoch": 1.13,
+ "learning_rate": 0.00048197115384615384,
+ "loss": 2.4557,
+ "step": 45
+ },
+ {
+ "epoch": 1.16,
+ "learning_rate": 0.00048157051282051285,
+ "loss": 2.4617,
+ "step": 46
+ },
+ {
+ "epoch": 1.18,
+ "learning_rate": 0.0004811698717948718,
+ "loss": 2.5267,
+ "step": 47
+ },
+ {
+ "epoch": 1.21,
+ "learning_rate": 0.0004807692307692308,
+ "loss": 2.4555,
+ "step": 48
+ },
+ {
+ "epoch": 1.23,
+ "learning_rate": 0.00048036858974358974,
+ "loss": 2.37,
+ "step": 49
+ },
+ {
+ "epoch": 1.26,
+ "learning_rate": 0.00047996794871794875,
+ "loss": 2.4829,
+ "step": 50
+ },
+ {
+ "epoch": 1.29,
+ "learning_rate": 0.0004795673076923077,
+ "loss": 2.5022,
+ "step": 51
+ },
+ {
+ "epoch": 1.31,
+ "learning_rate": 0.0004791666666666667,
+ "loss": 2.4958,
+ "step": 52
+ },
+ {
+ "epoch": 1.34,
+ "learning_rate": 0.0004787660256410257,
+ "loss": 2.4358,
+ "step": 53
+ },
+ {
+ "epoch": 1.36,
+ "learning_rate": 0.00047836538461538464,
+ "loss": 2.4742,
+ "step": 54
+ },
+ {
+ "epoch": 1.39,
+ "learning_rate": 0.00047796474358974365,
+ "loss": 2.4158,
+ "step": 55
+ },
+ {
+ "epoch": 1.41,
+ "learning_rate": 0.0004775641025641026,
+ "loss": 2.5216,
+ "step": 56
+ },
+ {
+ "epoch": 1.44,
+ "learning_rate": 0.0004771634615384616,
+ "loss": 2.5239,
+ "step": 57
+ },
+ {
+ "epoch": 1.46,
+ "learning_rate": 0.00047676282051282054,
+ "loss": 2.3946,
+ "step": 58
+ },
+ {
+ "epoch": 1.49,
+ "learning_rate": 0.0004763621794871795,
+ "loss": 2.4457,
+ "step": 59
+ },
+ {
+ "epoch": 1.51,
+ "learning_rate": 0.00047596153846153844,
+ "loss": 2.4379,
+ "step": 60
+ },
+ {
+ "epoch": 1.54,
+ "learning_rate": 0.00047556089743589744,
+ "loss": 2.4812,
+ "step": 61
+ },
+ {
+ "epoch": 1.56,
+ "learning_rate": 0.0004751602564102564,
+ "loss": 2.4153,
+ "step": 62
+ },
+ {
+ "epoch": 1.59,
+ "learning_rate": 0.0004747596153846154,
+ "loss": 2.4115,
+ "step": 63
+ },
+ {
+ "epoch": 1.61,
+ "learning_rate": 0.00047435897435897434,
+ "loss": 2.3637,
+ "step": 64
+ },
+ {
+ "epoch": 1.64,
+ "learning_rate": 0.00047395833333333334,
+ "loss": 2.3741,
+ "step": 65
+ },
+ {
+ "epoch": 1.66,
+ "learning_rate": 0.0004735576923076923,
+ "loss": 2.4461,
+ "step": 66
+ },
+ {
+ "epoch": 1.69,
+ "learning_rate": 0.0004731570512820513,
+ "loss": 2.4694,
+ "step": 67
+ },
+ {
+ "epoch": 1.71,
+ "learning_rate": 0.00047275641025641024,
+ "loss": 2.5177,
+ "step": 68
+ },
+ {
+ "epoch": 1.74,
+ "learning_rate": 0.00047235576923076924,
+ "loss": 2.4964,
+ "step": 69
+ },
+ {
+ "epoch": 1.76,
+ "learning_rate": 0.0004719551282051282,
+ "loss": 2.5114,
+ "step": 70
+ },
+ {
+ "epoch": 1.79,
+ "learning_rate": 0.0004715544871794872,
+ "loss": 2.4447,
+ "step": 71
+ },
+ {
+ "epoch": 1.81,
+ "learning_rate": 0.00047115384615384613,
+ "loss": 2.4112,
+ "step": 72
+ },
+ {
+ "epoch": 1.84,
+ "learning_rate": 0.00047075320512820514,
+ "loss": 2.4585,
+ "step": 73
+ },
+ {
+ "epoch": 1.86,
+ "learning_rate": 0.0004703525641025641,
+ "loss": 2.4696,
+ "step": 74
+ },
+ {
+ "epoch": 1.89,
+ "learning_rate": 0.0004699519230769231,
+ "loss": 2.3813,
+ "step": 75
+ },
+ {
+ "epoch": 1.91,
+ "learning_rate": 0.00046955128205128203,
+ "loss": 2.4962,
+ "step": 76
+ },
+ {
+ "epoch": 1.94,
+ "learning_rate": 0.00046915064102564104,
+ "loss": 2.4392,
+ "step": 77
+ },
+ {
+ "epoch": 1.97,
+ "learning_rate": 0.00046875,
+ "loss": 2.3436,
+ "step": 78
+ },
+ {
+ "epoch": 1.99,
+ "learning_rate": 0.000468349358974359,
+ "loss": 2.4758,
+ "step": 79
+ },
+ {
+ "epoch": 2.02,
+ "learning_rate": 0.000467948717948718,
+ "loss": 2.4026,
+ "step": 80
+ },
+ {
+ "epoch": 2.04,
+ "learning_rate": 0.00046754807692307693,
+ "loss": 2.3728,
+ "step": 81
+ },
+ {
+ "epoch": 2.07,
+ "learning_rate": 0.00046714743589743594,
+ "loss": 2.395,
+ "step": 82
+ },
+ {
+ "epoch": 2.09,
+ "learning_rate": 0.0004667467948717949,
+ "loss": 2.2767,
+ "step": 83
+ },
+ {
+ "epoch": 2.12,
+ "learning_rate": 0.0004663461538461539,
+ "loss": 2.3483,
+ "step": 84
+ },
+ {
+ "epoch": 2.14,
+ "learning_rate": 0.00046594551282051283,
+ "loss": 2.2911,
+ "step": 85
+ },
+ {
+ "epoch": 2.17,
+ "learning_rate": 0.00046554487179487183,
+ "loss": 2.3502,
+ "step": 86
+ },
+ {
+ "epoch": 2.19,
+ "learning_rate": 0.0004651442307692308,
+ "loss": 2.3532,
+ "step": 87
+ },
+ {
+ "epoch": 2.22,
+ "learning_rate": 0.0004647435897435898,
+ "loss": 2.266,
+ "step": 88
+ },
+ {
+ "epoch": 2.24,
+ "learning_rate": 0.00046434294871794873,
+ "loss": 2.3046,
+ "step": 89
+ },
+ {
+ "epoch": 2.27,
+ "learning_rate": 0.00046394230769230773,
+ "loss": 2.2458,
+ "step": 90
+ },
+ {
+ "epoch": 2.29,
+ "learning_rate": 0.0004635416666666667,
+ "loss": 2.3324,
+ "step": 91
+ },
+ {
+ "epoch": 2.32,
+ "learning_rate": 0.0004631410256410257,
+ "loss": 2.3016,
+ "step": 92
+ },
+ {
+ "epoch": 2.34,
+ "learning_rate": 0.00046274038461538463,
+ "loss": 2.379,
+ "step": 93
+ },
+ {
+ "epoch": 2.37,
+ "learning_rate": 0.00046233974358974363,
+ "loss": 2.3075,
+ "step": 94
+ },
+ {
+ "epoch": 2.39,
+ "learning_rate": 0.0004619391025641026,
+ "loss": 2.3942,
+ "step": 95
+ },
+ {
+ "epoch": 2.42,
+ "learning_rate": 0.0004615384615384616,
+ "loss": 2.2891,
+ "step": 96
+ },
+ {
+ "epoch": 2.44,
+ "learning_rate": 0.00046113782051282053,
+ "loss": 2.4199,
+ "step": 97
+ },
+ {
+ "epoch": 2.47,
+ "learning_rate": 0.0004607371794871795,
+ "loss": 2.3662,
+ "step": 98
+ },
+ {
+ "epoch": 2.49,
+ "learning_rate": 0.0004603365384615384,
+ "loss": 2.3346,
+ "step": 99
+ },
+ {
+ "epoch": 2.52,
+ "learning_rate": 0.0004599358974358974,
+ "loss": 2.2607,
+ "step": 100
+ },
+ {
+ "epoch": 2.54,
+ "learning_rate": 0.0004595352564102564,
+ "loss": 2.3446,
+ "step": 101
+ },
+ {
+ "epoch": 2.57,
+ "learning_rate": 0.0004591346153846154,
+ "loss": 2.3649,
+ "step": 102
+ },
+ {
+ "epoch": 2.6,
+ "learning_rate": 0.0004587339743589743,
+ "loss": 2.2394,
+ "step": 103
+ },
+ {
+ "epoch": 2.62,
+ "learning_rate": 0.0004583333333333333,
+ "loss": 2.3278,
+ "step": 104
+ },
+ {
+ "epoch": 2.65,
+ "learning_rate": 0.0004579326923076923,
+ "loss": 2.2725,
+ "step": 105
+ },
+ {
+ "epoch": 2.67,
+ "learning_rate": 0.0004575320512820513,
+ "loss": 2.3328,
+ "step": 106
+ },
+ {
+ "epoch": 2.7,
+ "learning_rate": 0.0004571314102564103,
+ "loss": 2.296,
+ "step": 107
+ },
+ {
+ "epoch": 2.72,
+ "learning_rate": 0.0004567307692307692,
+ "loss": 2.2562,
+ "step": 108
+ },
+ {
+ "epoch": 2.75,
+ "learning_rate": 0.0004563301282051282,
+ "loss": 2.3062,
+ "step": 109
+ },
+ {
+ "epoch": 2.77,
+ "learning_rate": 0.0004559294871794872,
+ "loss": 2.3125,
+ "step": 110
+ },
+ {
+ "epoch": 2.8,
+ "learning_rate": 0.0004555288461538462,
+ "loss": 2.3257,
+ "step": 111
+ },
+ {
+ "epoch": 2.82,
+ "learning_rate": 0.0004551282051282051,
+ "loss": 2.3431,
+ "step": 112
+ },
+ {
+ "epoch": 2.85,
+ "learning_rate": 0.0004547275641025641,
+ "loss": 2.3507,
+ "step": 113
+ },
+ {
+ "epoch": 2.87,
+ "learning_rate": 0.00045432692307692307,
+ "loss": 2.3621,
+ "step": 114
+ },
+ {
+ "epoch": 2.9,
+ "learning_rate": 0.0004539262820512821,
+ "loss": 2.2937,
+ "step": 115
+ },
+ {
+ "epoch": 2.92,
+ "learning_rate": 0.000453525641025641,
+ "loss": 2.3356,
+ "step": 116
+ },
+ {
+ "epoch": 2.95,
+ "learning_rate": 0.000453125,
+ "loss": 2.2228,
+ "step": 117
+ },
+ {
+ "epoch": 2.97,
+ "learning_rate": 0.00045272435897435897,
+ "loss": 2.3189,
+ "step": 118
+ },
+ {
+ "epoch": 3.0,
+ "learning_rate": 0.00045232371794871797,
+ "loss": 2.3605,
+ "step": 119
+ },
+ {
+ "epoch": 3.02,
+ "learning_rate": 0.0004519230769230769,
+ "loss": 2.1857,
+ "step": 120
+ },
+ {
+ "epoch": 3.05,
+ "learning_rate": 0.0004515224358974359,
+ "loss": 2.1538,
+ "step": 121
+ },
+ {
+ "epoch": 3.07,
+ "learning_rate": 0.00045112179487179487,
+ "loss": 2.2092,
+ "step": 122
+ },
+ {
+ "epoch": 3.1,
+ "learning_rate": 0.00045072115384615387,
+ "loss": 2.2089,
+ "step": 123
+ },
+ {
+ "epoch": 3.12,
+ "learning_rate": 0.0004503205128205128,
+ "loss": 2.1746,
+ "step": 124
+ },
+ {
+ "epoch": 3.15,
+ "learning_rate": 0.0004499198717948718,
+ "loss": 2.1455,
+ "step": 125
+ },
+ {
+ "epoch": 3.17,
+ "learning_rate": 0.00044951923076923077,
+ "loss": 2.119,
+ "step": 126
+ },
+ {
+ "epoch": 3.2,
+ "learning_rate": 0.00044911858974358977,
+ "loss": 2.2341,
+ "step": 127
+ },
+ {
+ "epoch": 3.23,
+ "learning_rate": 0.0004487179487179487,
+ "loss": 2.2275,
+ "step": 128
+ },
+ {
+ "epoch": 3.25,
+ "learning_rate": 0.0004483173076923077,
+ "loss": 2.176,
+ "step": 129
+ },
+ {
+ "epoch": 3.28,
+ "learning_rate": 0.0004479166666666667,
+ "loss": 2.1187,
+ "step": 130
+ },
+ {
+ "epoch": 3.3,
+ "learning_rate": 0.00044751602564102567,
+ "loss": 2.1175,
+ "step": 131
+ },
+ {
+ "epoch": 3.33,
+ "learning_rate": 0.00044711538461538467,
+ "loss": 2.1818,
+ "step": 132
+ },
+ {
+ "epoch": 3.35,
+ "learning_rate": 0.0004467147435897436,
+ "loss": 2.1193,
+ "step": 133
+ },
+ {
+ "epoch": 3.38,
+ "learning_rate": 0.0004463141025641026,
+ "loss": 2.2079,
+ "step": 134
+ },
+ {
+ "epoch": 3.4,
+ "learning_rate": 0.00044591346153846157,
+ "loss": 2.1361,
+ "step": 135
+ },
+ {
+ "epoch": 3.43,
+ "learning_rate": 0.00044551282051282057,
+ "loss": 2.2246,
+ "step": 136
+ },
+ {
+ "epoch": 3.45,
+ "learning_rate": 0.00044511217948717946,
+ "loss": 2.1139,
+ "step": 137
+ },
+ {
+ "epoch": 3.48,
+ "learning_rate": 0.00044471153846153846,
+ "loss": 2.1058,
+ "step": 138
+ },
+ {
+ "epoch": 3.5,
+ "learning_rate": 0.0004443108974358974,
+ "loss": 2.1442,
+ "step": 139
+ },
+ {
+ "epoch": 3.53,
+ "learning_rate": 0.0004439102564102564,
+ "loss": 2.1039,
+ "step": 140
+ },
+ {
+ "epoch": 3.55,
+ "learning_rate": 0.00044350961538461536,
+ "loss": 2.1884,
+ "step": 141
+ },
+ {
+ "epoch": 3.58,
+ "learning_rate": 0.00044310897435897436,
+ "loss": 2.21,
+ "step": 142
+ },
+ {
+ "epoch": 3.6,
+ "learning_rate": 0.0004427083333333333,
+ "loss": 2.1139,
+ "step": 143
+ },
+ {
+ "epoch": 3.63,
+ "learning_rate": 0.0004423076923076923,
+ "loss": 2.1049,
+ "step": 144
+ },
+ {
+ "epoch": 3.65,
+ "learning_rate": 0.00044190705128205126,
+ "loss": 2.1282,
+ "step": 145
+ },
+ {
+ "epoch": 3.68,
+ "learning_rate": 0.00044150641025641026,
+ "loss": 2.1803,
+ "step": 146
+ },
+ {
+ "epoch": 3.7,
+ "learning_rate": 0.0004411057692307692,
+ "loss": 2.1406,
+ "step": 147
+ },
+ {
+ "epoch": 3.73,
+ "learning_rate": 0.0004407051282051282,
+ "loss": 2.1841,
+ "step": 148
+ },
+ {
+ "epoch": 3.75,
+ "learning_rate": 0.00044030448717948716,
+ "loss": 2.1961,
+ "step": 149
+ },
+ {
+ "epoch": 3.78,
+ "learning_rate": 0.00043990384615384616,
+ "loss": 2.1265,
+ "step": 150
+ },
+ {
+ "epoch": 3.8,
+ "learning_rate": 0.0004395032051282051,
+ "loss": 2.1708,
+ "step": 151
+ },
+ {
+ "epoch": 3.83,
+ "learning_rate": 0.0004391025641025641,
+ "loss": 2.1555,
+ "step": 152
+ },
+ {
+ "epoch": 3.86,
+ "learning_rate": 0.00043870192307692306,
+ "loss": 2.1463,
+ "step": 153
+ },
+ {
+ "epoch": 3.88,
+ "learning_rate": 0.00043830128205128206,
+ "loss": 2.1947,
+ "step": 154
+ },
+ {
+ "epoch": 3.91,
+ "learning_rate": 0.000437900641025641,
+ "loss": 2.1962,
+ "step": 155
+ },
+ {
+ "epoch": 3.93,
+ "learning_rate": 0.0004375,
+ "loss": 2.2109,
+ "step": 156
+ },
+ {
+ "epoch": 3.96,
+ "learning_rate": 0.000437099358974359,
+ "loss": 2.1556,
+ "step": 157
+ },
+ {
+ "epoch": 3.98,
+ "learning_rate": 0.00043669871794871796,
+ "loss": 2.1605,
+ "step": 158
+ },
+ {
+ "epoch": 4.01,
+ "learning_rate": 0.00043629807692307696,
+ "loss": 2.155,
+ "step": 159
+ },
+ {
+ "epoch": 4.03,
+ "learning_rate": 0.0004358974358974359,
+ "loss": 2.0347,
+ "step": 160
+ },
+ {
+ "epoch": 4.06,
+ "learning_rate": 0.0004354967948717949,
+ "loss": 1.9321,
+ "step": 161
+ },
+ {
+ "epoch": 4.08,
+ "learning_rate": 0.00043509615384615386,
+ "loss": 2.0034,
+ "step": 162
+ },
+ {
+ "epoch": 4.11,
+ "learning_rate": 0.00043469551282051286,
+ "loss": 2.0203,
+ "step": 163
+ },
+ {
+ "epoch": 4.13,
+ "learning_rate": 0.0004342948717948718,
+ "loss": 1.9189,
+ "step": 164
+ },
+ {
+ "epoch": 4.16,
+ "learning_rate": 0.0004338942307692308,
+ "loss": 1.9868,
+ "step": 165
+ },
+ {
+ "epoch": 4.18,
+ "learning_rate": 0.00043349358974358976,
+ "loss": 1.9343,
+ "step": 166
+ },
+ {
+ "epoch": 4.21,
+ "learning_rate": 0.00043309294871794876,
+ "loss": 1.8894,
+ "step": 167
+ },
+ {
+ "epoch": 4.23,
+ "learning_rate": 0.0004326923076923077,
+ "loss": 1.9735,
+ "step": 168
+ },
+ {
+ "epoch": 4.26,
+ "learning_rate": 0.0004322916666666667,
+ "loss": 2.0012,
+ "step": 169
+ },
+ {
+ "epoch": 4.28,
+ "learning_rate": 0.00043189102564102565,
+ "loss": 1.9087,
+ "step": 170
+ },
+ {
+ "epoch": 4.31,
+ "learning_rate": 0.00043149038461538466,
+ "loss": 1.9213,
+ "step": 171
+ },
+ {
+ "epoch": 4.33,
+ "learning_rate": 0.0004310897435897436,
+ "loss": 2.0195,
+ "step": 172
+ },
+ {
+ "epoch": 4.36,
+ "learning_rate": 0.0004306891025641026,
+ "loss": 1.9806,
+ "step": 173
+ },
+ {
+ "epoch": 4.38,
+ "learning_rate": 0.00043028846153846155,
+ "loss": 1.9731,
+ "step": 174
+ },
+ {
+ "epoch": 4.41,
+ "learning_rate": 0.00042988782051282056,
+ "loss": 1.9282,
+ "step": 175
+ },
+ {
+ "epoch": 4.43,
+ "learning_rate": 0.00042948717948717945,
+ "loss": 1.9433,
+ "step": 176
+ },
+ {
+ "epoch": 4.46,
+ "learning_rate": 0.00042908653846153845,
+ "loss": 1.9633,
+ "step": 177
+ },
+ {
+ "epoch": 4.49,
+ "learning_rate": 0.0004286858974358974,
+ "loss": 1.9918,
+ "step": 178
+ },
+ {
+ "epoch": 4.51,
+ "learning_rate": 0.0004282852564102564,
+ "loss": 1.9265,
+ "step": 179
+ },
+ {
+ "epoch": 4.54,
+ "learning_rate": 0.00042788461538461535,
+ "loss": 1.9198,
+ "step": 180
+ },
+ {
+ "epoch": 4.56,
+ "learning_rate": 0.00042748397435897435,
+ "loss": 2.0159,
+ "step": 181
+ },
+ {
+ "epoch": 4.59,
+ "learning_rate": 0.0004270833333333333,
+ "loss": 1.9474,
+ "step": 182
+ },
+ {
+ "epoch": 4.61,
+ "learning_rate": 0.0004266826923076923,
+ "loss": 2.0152,
+ "step": 183
+ },
+ {
+ "epoch": 4.64,
+ "learning_rate": 0.0004262820512820513,
+ "loss": 1.9006,
+ "step": 184
+ },
+ {
+ "epoch": 4.66,
+ "learning_rate": 0.00042588141025641025,
+ "loss": 1.9862,
+ "step": 185
+ },
+ {
+ "epoch": 4.69,
+ "learning_rate": 0.00042548076923076925,
+ "loss": 1.9945,
+ "step": 186
+ },
+ {
+ "epoch": 4.71,
+ "learning_rate": 0.0004250801282051282,
+ "loss": 1.985,
+ "step": 187
+ },
+ {
+ "epoch": 4.74,
+ "learning_rate": 0.0004246794871794872,
+ "loss": 1.9899,
+ "step": 188
+ },
+ {
+ "epoch": 4.76,
+ "learning_rate": 0.00042427884615384615,
+ "loss": 2.0149,
+ "step": 189
+ },
+ {
+ "epoch": 4.79,
+ "learning_rate": 0.00042387820512820515,
+ "loss": 1.9768,
+ "step": 190
+ },
+ {
+ "epoch": 4.81,
+ "learning_rate": 0.0004234775641025641,
+ "loss": 2.0204,
+ "step": 191
+ },
+ {
+ "epoch": 4.84,
+ "learning_rate": 0.0004230769230769231,
+ "loss": 1.9343,
+ "step": 192
+ },
+ {
+ "epoch": 4.86,
+ "learning_rate": 0.00042267628205128205,
+ "loss": 2.0574,
+ "step": 193
+ },
+ {
+ "epoch": 4.89,
+ "learning_rate": 0.00042227564102564105,
+ "loss": 2.0245,
+ "step": 194
+ },
+ {
+ "epoch": 4.91,
+ "learning_rate": 0.000421875,
+ "loss": 2.0356,
+ "step": 195
+ },
+ {
+ "epoch": 4.94,
+ "learning_rate": 0.000421474358974359,
+ "loss": 1.9984,
+ "step": 196
+ },
+ {
+ "epoch": 4.96,
+ "learning_rate": 0.00042107371794871794,
+ "loss": 1.9826,
+ "step": 197
+ },
+ {
+ "epoch": 4.99,
+ "learning_rate": 0.00042067307692307695,
+ "loss": 1.9846,
+ "step": 198
+ },
+ {
+ "epoch": 5.01,
+ "learning_rate": 0.0004202724358974359,
+ "loss": 1.8931,
+ "step": 199
+ },
+ {
+ "epoch": 5.04,
+ "learning_rate": 0.0004198717948717949,
+ "loss": 1.7642,
+ "step": 200
+ },
+ {
+ "epoch": 5.06,
+ "learning_rate": 0.00041947115384615384,
+ "loss": 1.7235,
+ "step": 201
+ },
+ {
+ "epoch": 5.09,
+ "learning_rate": 0.00041907051282051285,
+ "loss": 1.7058,
+ "step": 202
+ },
+ {
+ "epoch": 5.11,
+ "learning_rate": 0.0004186698717948718,
+ "loss": 1.8022,
+ "step": 203
+ },
+ {
+ "epoch": 5.14,
+ "learning_rate": 0.0004182692307692308,
+ "loss": 1.7819,
+ "step": 204
+ },
+ {
+ "epoch": 5.17,
+ "learning_rate": 0.00041786858974358974,
+ "loss": 1.7047,
+ "step": 205
+ },
+ {
+ "epoch": 5.19,
+ "learning_rate": 0.00041746794871794874,
+ "loss": 1.773,
+ "step": 206
+ },
+ {
+ "epoch": 5.22,
+ "learning_rate": 0.0004170673076923077,
+ "loss": 1.7324,
+ "step": 207
+ },
+ {
+ "epoch": 5.24,
+ "learning_rate": 0.0004166666666666667,
+ "loss": 1.7862,
+ "step": 208
+ },
+ {
+ "epoch": 5.27,
+ "learning_rate": 0.0004162660256410257,
+ "loss": 1.764,
+ "step": 209
+ },
+ {
+ "epoch": 5.29,
+ "learning_rate": 0.00041586538461538464,
+ "loss": 1.7156,
+ "step": 210
+ },
+ {
+ "epoch": 5.32,
+ "learning_rate": 0.00041546474358974364,
+ "loss": 1.7948,
+ "step": 211
+ },
+ {
+ "epoch": 5.34,
+ "learning_rate": 0.0004150641025641026,
+ "loss": 1.7608,
+ "step": 212
+ },
+ {
+ "epoch": 5.37,
+ "learning_rate": 0.0004146634615384616,
+ "loss": 1.7204,
+ "step": 213
+ },
+ {
+ "epoch": 5.39,
+ "learning_rate": 0.00041426282051282054,
+ "loss": 1.7691,
+ "step": 214
+ },
+ {
+ "epoch": 5.42,
+ "learning_rate": 0.0004138621794871795,
+ "loss": 1.7969,
+ "step": 215
+ },
+ {
+ "epoch": 5.44,
+ "learning_rate": 0.00041346153846153844,
+ "loss": 1.7194,
+ "step": 216
+ },
+ {
+ "epoch": 5.47,
+ "learning_rate": 0.00041306089743589744,
+ "loss": 1.7022,
+ "step": 217
+ },
+ {
+ "epoch": 5.49,
+ "learning_rate": 0.0004126602564102564,
+ "loss": 1.8287,
+ "step": 218
+ },
+ {
+ "epoch": 5.52,
+ "learning_rate": 0.0004122596153846154,
+ "loss": 1.7074,
+ "step": 219
+ },
+ {
+ "epoch": 5.54,
+ "learning_rate": 0.00041185897435897434,
+ "loss": 1.8354,
+ "step": 220
+ },
+ {
+ "epoch": 5.57,
+ "learning_rate": 0.00041145833333333334,
+ "loss": 1.791,
+ "step": 221
+ },
+ {
+ "epoch": 5.59,
+ "learning_rate": 0.0004110576923076923,
+ "loss": 1.7633,
+ "step": 222
+ },
+ {
+ "epoch": 5.62,
+ "learning_rate": 0.0004106570512820513,
+ "loss": 1.811,
+ "step": 223
+ },
+ {
+ "epoch": 5.64,
+ "learning_rate": 0.00041025641025641023,
+ "loss": 1.7498,
+ "step": 224
+ },
+ {
+ "epoch": 5.67,
+ "learning_rate": 0.00040985576923076924,
+ "loss": 1.7662,
+ "step": 225
+ },
+ {
+ "epoch": 5.69,
+ "learning_rate": 0.0004094551282051282,
+ "loss": 1.7686,
+ "step": 226
+ },
+ {
+ "epoch": 5.72,
+ "learning_rate": 0.0004090544871794872,
+ "loss": 1.8357,
+ "step": 227
+ },
+ {
+ "epoch": 5.74,
+ "learning_rate": 0.00040865384615384613,
+ "loss": 1.7981,
+ "step": 228
+ },
+ {
+ "epoch": 5.77,
+ "learning_rate": 0.00040825320512820513,
+ "loss": 1.7502,
+ "step": 229
+ },
+ {
+ "epoch": 5.8,
+ "learning_rate": 0.0004078525641025641,
+ "loss": 1.7666,
+ "step": 230
+ },
+ {
+ "epoch": 5.82,
+ "learning_rate": 0.0004074519230769231,
+ "loss": 1.798,
+ "step": 231
+ },
+ {
+ "epoch": 5.85,
+ "learning_rate": 0.00040705128205128203,
+ "loss": 1.8072,
+ "step": 232
+ },
+ {
+ "epoch": 5.87,
+ "learning_rate": 0.00040665064102564103,
+ "loss": 1.7611,
+ "step": 233
+ },
+ {
+ "epoch": 5.9,
+ "learning_rate": 0.00040625000000000004,
+ "loss": 1.8126,
+ "step": 234
+ },
+ {
+ "epoch": 5.92,
+ "learning_rate": 0.000405849358974359,
+ "loss": 1.8361,
+ "step": 235
+ },
+ {
+ "epoch": 5.95,
+ "learning_rate": 0.000405448717948718,
+ "loss": 1.7938,
+ "step": 236
+ },
+ {
+ "epoch": 5.97,
+ "learning_rate": 0.00040504807692307693,
+ "loss": 1.7524,
+ "step": 237
+ },
+ {
+ "epoch": 6.0,
+ "learning_rate": 0.00040464743589743593,
+ "loss": 1.7601,
+ "step": 238
+ },
+ {
+ "epoch": 6.02,
+ "learning_rate": 0.0004042467948717949,
+ "loss": 1.5373,
+ "step": 239
+ },
+ {
+ "epoch": 6.05,
+ "learning_rate": 0.0004038461538461539,
+ "loss": 1.5108,
+ "step": 240
+ },
+ {
+ "epoch": 6.07,
+ "learning_rate": 0.00040344551282051283,
+ "loss": 1.568,
+ "step": 241
+ },
+ {
+ "epoch": 6.1,
+ "learning_rate": 0.00040304487179487183,
+ "loss": 1.5175,
+ "step": 242
+ },
+ {
+ "epoch": 6.12,
+ "learning_rate": 0.0004026442307692308,
+ "loss": 1.5748,
+ "step": 243
+ },
+ {
+ "epoch": 6.15,
+ "learning_rate": 0.0004022435897435898,
+ "loss": 1.4772,
+ "step": 244
+ },
+ {
+ "epoch": 6.17,
+ "learning_rate": 0.00040184294871794873,
+ "loss": 1.5035,
+ "step": 245
+ },
+ {
+ "epoch": 6.2,
+ "learning_rate": 0.00040144230769230773,
+ "loss": 1.5318,
+ "step": 246
+ },
+ {
+ "epoch": 6.22,
+ "learning_rate": 0.0004010416666666667,
+ "loss": 1.5139,
+ "step": 247
+ },
+ {
+ "epoch": 6.25,
+ "learning_rate": 0.0004006410256410257,
+ "loss": 1.5599,
+ "step": 248
+ },
+ {
+ "epoch": 6.27,
+ "learning_rate": 0.00040024038461538463,
+ "loss": 1.5475,
+ "step": 249
+ },
+ {
+ "epoch": 6.3,
+ "learning_rate": 0.00039983974358974363,
+ "loss": 1.4819,
+ "step": 250
+ },
+ {
+ "epoch": 6.32,
+ "learning_rate": 0.0003994391025641026,
+ "loss": 1.5419,
+ "step": 251
+ },
+ {
+ "epoch": 6.35,
+ "learning_rate": 0.0003990384615384616,
+ "loss": 1.6027,
+ "step": 252
+ },
+ {
+ "epoch": 6.37,
+ "learning_rate": 0.00039863782051282053,
+ "loss": 1.5683,
+ "step": 253
+ },
+ {
+ "epoch": 6.4,
+ "learning_rate": 0.0003982371794871795,
+ "loss": 1.5963,
+ "step": 254
+ },
+ {
+ "epoch": 6.43,
+ "learning_rate": 0.0003978365384615384,
+ "loss": 1.5874,
+ "step": 255
+ },
+ {
+ "epoch": 6.45,
+ "learning_rate": 0.0003974358974358974,
+ "loss": 1.5161,
+ "step": 256
+ },
+ {
+ "epoch": 6.48,
+ "learning_rate": 0.00039703525641025637,
+ "loss": 1.57,
+ "step": 257
+ },
+ {
+ "epoch": 6.5,
+ "learning_rate": 0.0003966346153846154,
+ "loss": 1.534,
+ "step": 258
+ },
+ {
+ "epoch": 6.53,
+ "learning_rate": 0.0003962339743589743,
+ "loss": 1.5513,
+ "step": 259
+ },
+ {
+ "epoch": 6.55,
+ "learning_rate": 0.0003958333333333333,
+ "loss": 1.5454,
+ "step": 260
+ },
+ {
+ "epoch": 6.58,
+ "learning_rate": 0.0003954326923076923,
+ "loss": 1.5695,
+ "step": 261
+ },
+ {
+ "epoch": 6.6,
+ "learning_rate": 0.0003950320512820513,
+ "loss": 1.5777,
+ "step": 262
+ },
+ {
+ "epoch": 6.63,
+ "learning_rate": 0.0003946314102564103,
+ "loss": 1.5629,
+ "step": 263
+ },
+ {
+ "epoch": 6.65,
+ "learning_rate": 0.0003942307692307692,
+ "loss": 1.5031,
+ "step": 264
+ },
+ {
+ "epoch": 6.68,
+ "learning_rate": 0.0003938301282051282,
+ "loss": 1.5964,
+ "step": 265
+ },
+ {
+ "epoch": 6.7,
+ "learning_rate": 0.00039342948717948717,
+ "loss": 1.5472,
+ "step": 266
+ },
+ {
+ "epoch": 6.73,
+ "learning_rate": 0.0003930288461538462,
+ "loss": 1.5654,
+ "step": 267
+ },
+ {
+ "epoch": 6.75,
+ "learning_rate": 0.0003926282051282051,
+ "loss": 1.5932,
+ "step": 268
+ },
+ {
+ "epoch": 6.78,
+ "learning_rate": 0.0003922275641025641,
+ "loss": 1.6037,
+ "step": 269
+ },
+ {
+ "epoch": 6.8,
+ "learning_rate": 0.00039182692307692307,
+ "loss": 1.6394,
+ "step": 270
+ },
+ {
+ "epoch": 6.83,
+ "learning_rate": 0.00039142628205128207,
+ "loss": 1.5745,
+ "step": 271
+ },
+ {
+ "epoch": 6.85,
+ "learning_rate": 0.000391025641025641,
+ "loss": 1.6512,
+ "step": 272
+ },
+ {
+ "epoch": 6.88,
+ "learning_rate": 0.000390625,
+ "loss": 1.6107,
+ "step": 273
+ },
+ {
+ "epoch": 6.9,
+ "learning_rate": 0.00039022435897435897,
+ "loss": 1.5798,
+ "step": 274
+ },
+ {
+ "epoch": 6.93,
+ "learning_rate": 0.00038982371794871797,
+ "loss": 1.6087,
+ "step": 275
+ },
+ {
+ "epoch": 6.95,
+ "learning_rate": 0.0003894230769230769,
+ "loss": 1.5362,
+ "step": 276
+ },
+ {
+ "epoch": 6.98,
+ "learning_rate": 0.0003890224358974359,
+ "loss": 1.5899,
+ "step": 277
+ },
+ {
+ "epoch": 7.0,
+ "learning_rate": 0.00038862179487179487,
+ "loss": 1.59,
+ "step": 278
+ },
+ {
+ "epoch": 7.03,
+ "learning_rate": 0.00038822115384615387,
+ "loss": 1.3327,
+ "step": 279
+ },
+ {
+ "epoch": 7.06,
+ "learning_rate": 0.0003878205128205128,
+ "loss": 1.3214,
+ "step": 280
+ },
+ {
+ "epoch": 7.08,
+ "learning_rate": 0.0003874198717948718,
+ "loss": 1.2891,
+ "step": 281
+ },
+ {
+ "epoch": 7.11,
+ "learning_rate": 0.00038701923076923077,
+ "loss": 1.3321,
+ "step": 282
+ },
+ {
+ "epoch": 7.13,
+ "learning_rate": 0.00038661858974358977,
+ "loss": 1.284,
+ "step": 283
+ },
+ {
+ "epoch": 7.16,
+ "learning_rate": 0.0003862179487179487,
+ "loss": 1.3346,
+ "step": 284
+ },
+ {
+ "epoch": 7.18,
+ "learning_rate": 0.0003858173076923077,
+ "loss": 1.2874,
+ "step": 285
+ },
+ {
+ "epoch": 7.21,
+ "learning_rate": 0.0003854166666666667,
+ "loss": 1.3104,
+ "step": 286
+ },
+ {
+ "epoch": 7.23,
+ "learning_rate": 0.00038501602564102567,
+ "loss": 1.3138,
+ "step": 287
+ },
+ {
+ "epoch": 7.26,
+ "learning_rate": 0.00038461538461538467,
+ "loss": 1.2871,
+ "step": 288
+ },
+ {
+ "epoch": 7.28,
+ "learning_rate": 0.0003842147435897436,
+ "loss": 1.2885,
+ "step": 289
+ },
+ {
+ "epoch": 7.31,
+ "learning_rate": 0.0003838141025641026,
+ "loss": 1.338,
+ "step": 290
+ },
+ {
+ "epoch": 7.33,
+ "learning_rate": 0.00038341346153846157,
+ "loss": 1.3189,
+ "step": 291
+ },
+ {
+ "epoch": 7.36,
+ "learning_rate": 0.00038301282051282057,
+ "loss": 1.3013,
+ "step": 292
+ },
+ {
+ "epoch": 7.38,
+ "learning_rate": 0.00038261217948717946,
+ "loss": 1.3505,
+ "step": 293
+ },
+ {
+ "epoch": 7.41,
+ "learning_rate": 0.00038221153846153846,
+ "loss": 1.3469,
+ "step": 294
+ },
+ {
+ "epoch": 7.43,
+ "learning_rate": 0.0003818108974358974,
+ "loss": 1.3156,
+ "step": 295
+ },
+ {
+ "epoch": 7.46,
+ "learning_rate": 0.0003814102564102564,
+ "loss": 1.3796,
+ "step": 296
+ },
+ {
+ "epoch": 7.48,
+ "learning_rate": 0.00038100961538461536,
+ "loss": 1.3649,
+ "step": 297
+ },
+ {
+ "epoch": 7.51,
+ "learning_rate": 0.00038060897435897436,
+ "loss": 1.3483,
+ "step": 298
+ },
+ {
+ "epoch": 7.53,
+ "learning_rate": 0.0003802083333333333,
+ "loss": 1.3481,
+ "step": 299
+ },
+ {
+ "epoch": 7.56,
+ "learning_rate": 0.0003798076923076923,
+ "loss": 1.334,
+ "step": 300
+ },
+ {
+ "epoch": 7.58,
+ "learning_rate": 0.00037940705128205126,
+ "loss": 1.3495,
+ "step": 301
+ },
+ {
+ "epoch": 7.61,
+ "learning_rate": 0.00037900641025641026,
+ "loss": 1.4102,
+ "step": 302
+ },
+ {
+ "epoch": 7.63,
+ "learning_rate": 0.0003786057692307692,
+ "loss": 1.3794,
+ "step": 303
+ },
+ {
+ "epoch": 7.66,
+ "learning_rate": 0.0003782051282051282,
+ "loss": 1.3802,
+ "step": 304
+ },
+ {
+ "epoch": 7.69,
+ "learning_rate": 0.00037780448717948716,
+ "loss": 1.3835,
+ "step": 305
+ },
+ {
+ "epoch": 7.71,
+ "learning_rate": 0.00037740384615384616,
+ "loss": 1.4197,
+ "step": 306
+ },
+ {
+ "epoch": 7.74,
+ "learning_rate": 0.0003770032051282051,
+ "loss": 1.3947,
+ "step": 307
+ },
+ {
+ "epoch": 7.76,
+ "learning_rate": 0.0003766025641025641,
+ "loss": 1.3891,
+ "step": 308
+ },
+ {
+ "epoch": 7.79,
+ "learning_rate": 0.00037620192307692306,
+ "loss": 1.3904,
+ "step": 309
+ },
+ {
+ "epoch": 7.81,
+ "learning_rate": 0.00037580128205128206,
+ "loss": 1.3937,
+ "step": 310
+ },
+ {
+ "epoch": 7.84,
+ "learning_rate": 0.000375400641025641,
+ "loss": 1.4258,
+ "step": 311
+ },
+ {
+ "epoch": 7.86,
+ "learning_rate": 0.000375,
+ "loss": 1.3816,
+ "step": 312
+ },
+ {
+ "epoch": 7.89,
+ "learning_rate": 0.000374599358974359,
+ "loss": 1.4251,
+ "step": 313
+ },
+ {
+ "epoch": 7.91,
+ "learning_rate": 0.00037419871794871796,
+ "loss": 1.3946,
+ "step": 314
+ },
+ {
+ "epoch": 7.94,
+ "learning_rate": 0.00037379807692307696,
+ "loss": 1.3968,
+ "step": 315
+ },
+ {
+ "epoch": 7.96,
+ "learning_rate": 0.0003733974358974359,
+ "loss": 1.4108,
+ "step": 316
+ },
+ {
+ "epoch": 7.99,
+ "learning_rate": 0.0003729967948717949,
+ "loss": 1.4673,
+ "step": 317
+ },
+ {
+ "epoch": 8.01,
+ "learning_rate": 0.00037259615384615386,
+ "loss": 1.2339,
+ "step": 318
+ },
+ {
+ "epoch": 8.04,
+ "learning_rate": 0.00037219551282051286,
+ "loss": 1.1775,
+ "step": 319
+ },
+ {
+ "epoch": 8.06,
+ "learning_rate": 0.0003717948717948718,
+ "loss": 1.1061,
+ "step": 320
+ },
+ {
+ "epoch": 8.09,
+ "learning_rate": 0.0003713942307692308,
+ "loss": 1.0912,
+ "step": 321
+ },
+ {
+ "epoch": 8.11,
+ "learning_rate": 0.00037099358974358975,
+ "loss": 1.1214,
+ "step": 322
+ },
+ {
+ "epoch": 8.14,
+ "learning_rate": 0.00037059294871794876,
+ "loss": 1.1537,
+ "step": 323
+ },
+ {
+ "epoch": 8.16,
+ "learning_rate": 0.0003701923076923077,
+ "loss": 1.116,
+ "step": 324
+ },
+ {
+ "epoch": 8.19,
+ "learning_rate": 0.0003697916666666667,
+ "loss": 1.1292,
+ "step": 325
+ },
+ {
+ "epoch": 8.21,
+ "learning_rate": 0.00036939102564102565,
+ "loss": 1.0948,
+ "step": 326
+ },
+ {
+ "epoch": 8.24,
+ "learning_rate": 0.00036899038461538466,
+ "loss": 1.0968,
+ "step": 327
+ },
+ {
+ "epoch": 8.26,
+ "learning_rate": 0.0003685897435897436,
+ "loss": 1.118,
+ "step": 328
+ },
+ {
+ "epoch": 8.29,
+ "learning_rate": 0.0003681891025641026,
+ "loss": 1.1488,
+ "step": 329
+ },
+ {
+ "epoch": 8.31,
+ "learning_rate": 0.00036778846153846155,
+ "loss": 1.137,
+ "step": 330
+ },
+ {
+ "epoch": 8.34,
+ "learning_rate": 0.00036738782051282055,
+ "loss": 1.0832,
+ "step": 331
+ },
+ {
+ "epoch": 8.37,
+ "learning_rate": 0.00036698717948717945,
+ "loss": 1.1609,
+ "step": 332
+ },
+ {
+ "epoch": 8.39,
+ "learning_rate": 0.00036658653846153845,
+ "loss": 1.1497,
+ "step": 333
+ },
+ {
+ "epoch": 8.42,
+ "learning_rate": 0.0003661858974358974,
+ "loss": 1.1289,
+ "step": 334
+ },
+ {
+ "epoch": 8.44,
+ "learning_rate": 0.0003657852564102564,
+ "loss": 1.1938,
+ "step": 335
+ },
+ {
+ "epoch": 8.47,
+ "learning_rate": 0.00036538461538461535,
+ "loss": 1.1441,
+ "step": 336
+ },
+ {
+ "epoch": 8.49,
+ "learning_rate": 0.00036498397435897435,
+ "loss": 1.1233,
+ "step": 337
+ },
+ {
+ "epoch": 8.52,
+ "learning_rate": 0.0003645833333333333,
+ "loss": 1.1273,
+ "step": 338
+ },
+ {
+ "epoch": 8.54,
+ "learning_rate": 0.0003641826923076923,
+ "loss": 1.1498,
+ "step": 339
+ },
+ {
+ "epoch": 8.57,
+ "learning_rate": 0.0003637820512820513,
+ "loss": 1.1848,
+ "step": 340
+ },
+ {
+ "epoch": 8.59,
+ "learning_rate": 0.00036338141025641025,
+ "loss": 1.1991,
+ "step": 341
+ },
+ {
+ "epoch": 8.62,
+ "learning_rate": 0.00036298076923076925,
+ "loss": 1.1705,
+ "step": 342
+ },
+ {
+ "epoch": 8.64,
+ "learning_rate": 0.0003625801282051282,
+ "loss": 1.1613,
+ "step": 343
+ },
+ {
+ "epoch": 8.67,
+ "learning_rate": 0.0003621794871794872,
+ "loss": 1.2599,
+ "step": 344
+ },
+ {
+ "epoch": 8.69,
+ "learning_rate": 0.00036177884615384615,
+ "loss": 1.1824,
+ "step": 345
+ },
+ {
+ "epoch": 8.72,
+ "learning_rate": 0.00036137820512820515,
+ "loss": 1.1763,
+ "step": 346
+ },
+ {
+ "epoch": 8.74,
+ "learning_rate": 0.0003609775641025641,
+ "loss": 1.1882,
+ "step": 347
+ },
+ {
+ "epoch": 8.77,
+ "learning_rate": 0.0003605769230769231,
+ "loss": 1.2043,
+ "step": 348
+ },
+ {
+ "epoch": 8.79,
+ "learning_rate": 0.00036017628205128204,
+ "loss": 1.2123,
+ "step": 349
+ },
+ {
+ "epoch": 8.82,
+ "learning_rate": 0.00035977564102564105,
+ "loss": 1.1831,
+ "step": 350
+ },
+ {
+ "epoch": 8.84,
+ "learning_rate": 0.000359375,
+ "loss": 1.2265,
+ "step": 351
+ },
+ {
+ "epoch": 8.87,
+ "learning_rate": 0.000358974358974359,
+ "loss": 1.1658,
+ "step": 352
+ },
+ {
+ "epoch": 8.89,
+ "learning_rate": 0.00035857371794871794,
+ "loss": 1.1787,
+ "step": 353
+ },
+ {
+ "epoch": 8.92,
+ "learning_rate": 0.00035817307692307695,
+ "loss": 1.1904,
+ "step": 354
+ },
+ {
+ "epoch": 8.94,
+ "learning_rate": 0.0003577724358974359,
+ "loss": 1.2753,
+ "step": 355
+ },
+ {
+ "epoch": 8.97,
+ "learning_rate": 0.0003573717948717949,
+ "loss": 1.2076,
+ "step": 356
+ },
+ {
+ "epoch": 9.0,
+ "learning_rate": 0.00035697115384615384,
+ "loss": 1.2433,
+ "step": 357
+ },
+ {
+ "epoch": 9.02,
+ "learning_rate": 0.00035657051282051284,
+ "loss": 0.9733,
+ "step": 358
+ },
+ {
+ "epoch": 9.05,
+ "learning_rate": 0.0003561698717948718,
+ "loss": 0.9765,
+ "step": 359
+ },
+ {
+ "epoch": 9.07,
+ "learning_rate": 0.0003557692307692308,
+ "loss": 0.9626,
+ "step": 360
+ },
+ {
+ "epoch": 9.1,
+ "learning_rate": 0.00035536858974358974,
+ "loss": 0.9382,
+ "step": 361
+ },
+ {
+ "epoch": 9.12,
+ "learning_rate": 0.00035496794871794874,
+ "loss": 0.8644,
+ "step": 362
+ },
+ {
+ "epoch": 9.15,
+ "learning_rate": 0.0003545673076923077,
+ "loss": 0.9299,
+ "step": 363
+ },
+ {
+ "epoch": 9.17,
+ "learning_rate": 0.0003541666666666667,
+ "loss": 0.9618,
+ "step": 364
+ },
+ {
+ "epoch": 9.2,
+ "learning_rate": 0.0003537660256410257,
+ "loss": 0.9417,
+ "step": 365
+ },
+ {
+ "epoch": 9.22,
+ "learning_rate": 0.00035336538461538464,
+ "loss": 0.9873,
+ "step": 366
+ },
+ {
+ "epoch": 9.25,
+ "learning_rate": 0.00035296474358974364,
+ "loss": 0.9135,
+ "step": 367
+ },
+ {
+ "epoch": 9.27,
+ "learning_rate": 0.0003525641025641026,
+ "loss": 0.9652,
+ "step": 368
+ },
+ {
+ "epoch": 9.3,
+ "learning_rate": 0.0003521634615384616,
+ "loss": 0.9347,
+ "step": 369
+ },
+ {
+ "epoch": 9.32,
+ "learning_rate": 0.00035176282051282054,
+ "loss": 0.914,
+ "step": 370
+ },
+ {
+ "epoch": 9.35,
+ "learning_rate": 0.0003513621794871795,
+ "loss": 0.9622,
+ "step": 371
+ },
+ {
+ "epoch": 9.37,
+ "learning_rate": 0.00035096153846153844,
+ "loss": 0.9652,
+ "step": 372
+ },
+ {
+ "epoch": 9.4,
+ "learning_rate": 0.00035056089743589744,
+ "loss": 0.8865,
+ "step": 373
+ },
+ {
+ "epoch": 9.42,
+ "learning_rate": 0.0003501602564102564,
+ "loss": 0.9997,
+ "step": 374
+ },
+ {
+ "epoch": 9.45,
+ "learning_rate": 0.0003497596153846154,
+ "loss": 0.9695,
+ "step": 375
+ },
+ {
+ "epoch": 9.47,
+ "learning_rate": 0.00034935897435897433,
+ "loss": 1.0172,
+ "step": 376
+ },
+ {
+ "epoch": 9.5,
+ "learning_rate": 0.00034895833333333334,
+ "loss": 0.9955,
+ "step": 377
+ },
+ {
+ "epoch": 9.52,
+ "learning_rate": 0.0003485576923076923,
+ "loss": 0.9902,
+ "step": 378
+ },
+ {
+ "epoch": 9.55,
+ "learning_rate": 0.0003481570512820513,
+ "loss": 1.0225,
+ "step": 379
+ },
+ {
+ "epoch": 9.57,
+ "learning_rate": 0.00034775641025641023,
+ "loss": 0.9773,
+ "step": 380
+ },
+ {
+ "epoch": 9.6,
+ "learning_rate": 0.00034735576923076923,
+ "loss": 0.9947,
+ "step": 381
+ },
+ {
+ "epoch": 9.63,
+ "learning_rate": 0.0003469551282051282,
+ "loss": 1.0234,
+ "step": 382
+ },
+ {
+ "epoch": 9.65,
+ "learning_rate": 0.0003465544871794872,
+ "loss": 0.9692,
+ "step": 383
+ },
+ {
+ "epoch": 9.68,
+ "learning_rate": 0.00034615384615384613,
+ "loss": 0.9782,
+ "step": 384
+ },
+ {
+ "epoch": 9.7,
+ "learning_rate": 0.00034575320512820513,
+ "loss": 1.0416,
+ "step": 385
+ },
+ {
+ "epoch": 9.73,
+ "learning_rate": 0.0003453525641025641,
+ "loss": 1.0008,
+ "step": 386
+ },
+ {
+ "epoch": 9.75,
+ "learning_rate": 0.0003449519230769231,
+ "loss": 1.0854,
+ "step": 387
+ },
+ {
+ "epoch": 9.78,
+ "learning_rate": 0.00034455128205128203,
+ "loss": 1.085,
+ "step": 388
+ },
+ {
+ "epoch": 9.8,
+ "learning_rate": 0.00034415064102564103,
+ "loss": 1.007,
+ "step": 389
+ },
+ {
+ "epoch": 9.83,
+ "learning_rate": 0.00034375,
+ "loss": 1.0516,
+ "step": 390
+ },
+ {
+ "epoch": 9.85,
+ "learning_rate": 0.000343349358974359,
+ "loss": 0.9729,
+ "step": 391
+ },
+ {
+ "epoch": 9.88,
+ "learning_rate": 0.000342948717948718,
+ "loss": 1.0174,
+ "step": 392
+ },
+ {
+ "epoch": 9.9,
+ "learning_rate": 0.00034254807692307693,
+ "loss": 1.0392,
+ "step": 393
+ },
+ {
+ "epoch": 9.93,
+ "learning_rate": 0.00034214743589743593,
+ "loss": 1.0534,
+ "step": 394
+ },
+ {
+ "epoch": 9.95,
+ "learning_rate": 0.0003417467948717949,
+ "loss": 1.0584,
+ "step": 395
+ },
+ {
+ "epoch": 9.98,
+ "learning_rate": 0.0003413461538461539,
+ "loss": 1.0727,
+ "step": 396
+ },
+ {
+ "epoch": 10.0,
+ "learning_rate": 0.00034094551282051283,
+ "loss": 1.0023,
+ "step": 397
+ },
+ {
+ "epoch": 10.03,
+ "learning_rate": 0.00034054487179487183,
+ "loss": 0.8029,
+ "step": 398
+ },
+ {
+ "epoch": 10.05,
+ "learning_rate": 0.0003401442307692308,
+ "loss": 0.7901,
+ "step": 399
+ },
+ {
+ "epoch": 10.08,
+ "learning_rate": 0.0003397435897435898,
+ "loss": 0.7791,
+ "step": 400
+ },
+ {
+ "epoch": 10.1,
+ "learning_rate": 0.00033934294871794873,
+ "loss": 0.7635,
+ "step": 401
+ },
+ {
+ "epoch": 10.13,
+ "learning_rate": 0.00033894230769230773,
+ "loss": 0.7713,
+ "step": 402
+ },
+ {
+ "epoch": 10.15,
+ "learning_rate": 0.0003385416666666667,
+ "loss": 0.827,
+ "step": 403
+ },
+ {
+ "epoch": 10.18,
+ "learning_rate": 0.0003381410256410257,
+ "loss": 0.7844,
+ "step": 404
+ },
+ {
+ "epoch": 10.2,
+ "learning_rate": 0.00033774038461538463,
+ "loss": 0.7839,
+ "step": 405
+ },
+ {
+ "epoch": 10.23,
+ "learning_rate": 0.00033733974358974363,
+ "loss": 0.7702,
+ "step": 406
+ },
+ {
+ "epoch": 10.26,
+ "learning_rate": 0.0003369391025641026,
+ "loss": 0.7855,
+ "step": 407
+ },
+ {
+ "epoch": 10.28,
+ "learning_rate": 0.0003365384615384616,
+ "loss": 0.7669,
+ "step": 408
+ },
+ {
+ "epoch": 10.31,
+ "learning_rate": 0.0003361378205128205,
+ "loss": 0.8054,
+ "step": 409
+ },
+ {
+ "epoch": 10.33,
+ "learning_rate": 0.0003357371794871795,
+ "loss": 0.7792,
+ "step": 410
+ },
+ {
+ "epoch": 10.36,
+ "learning_rate": 0.0003353365384615384,
+ "loss": 0.7986,
+ "step": 411
+ },
+ {
+ "epoch": 10.38,
+ "learning_rate": 0.0003349358974358974,
+ "loss": 0.8269,
+ "step": 412
+ },
+ {
+ "epoch": 10.41,
+ "learning_rate": 0.00033453525641025637,
+ "loss": 0.7923,
+ "step": 413
+ },
+ {
+ "epoch": 10.43,
+ "learning_rate": 0.00033413461538461537,
+ "loss": 0.8242,
+ "step": 414
+ },
+ {
+ "epoch": 10.46,
+ "learning_rate": 0.0003337339743589743,
+ "loss": 0.8496,
+ "step": 415
+ },
+ {
+ "epoch": 10.48,
+ "learning_rate": 0.0003333333333333333,
+ "loss": 0.8236,
+ "step": 416
+ },
+ {
+ "epoch": 10.51,
+ "learning_rate": 0.0003329326923076923,
+ "loss": 0.8685,
+ "step": 417
+ },
+ {
+ "epoch": 10.53,
+ "learning_rate": 0.00033253205128205127,
+ "loss": 0.8542,
+ "step": 418
+ },
+ {
+ "epoch": 10.56,
+ "learning_rate": 0.0003321314102564103,
+ "loss": 0.8185,
+ "step": 419
+ },
+ {
+ "epoch": 10.58,
+ "learning_rate": 0.0003317307692307692,
+ "loss": 0.8477,
+ "step": 420
+ },
+ {
+ "epoch": 10.61,
+ "learning_rate": 0.0003313301282051282,
+ "loss": 0.8372,
+ "step": 421
+ },
+ {
+ "epoch": 10.63,
+ "learning_rate": 0.00033092948717948717,
+ "loss": 0.8731,
+ "step": 422
+ },
+ {
+ "epoch": 10.66,
+ "learning_rate": 0.00033052884615384617,
+ "loss": 0.8322,
+ "step": 423
+ },
+ {
+ "epoch": 10.68,
+ "learning_rate": 0.0003301282051282051,
+ "loss": 0.8514,
+ "step": 424
+ },
+ {
+ "epoch": 10.71,
+ "learning_rate": 0.0003297275641025641,
+ "loss": 0.8564,
+ "step": 425
+ },
+ {
+ "epoch": 10.73,
+ "learning_rate": 0.00032932692307692307,
+ "loss": 0.8278,
+ "step": 426
+ },
+ {
+ "epoch": 10.76,
+ "learning_rate": 0.00032892628205128207,
+ "loss": 0.8408,
+ "step": 427
+ },
+ {
+ "epoch": 10.78,
+ "learning_rate": 0.000328525641025641,
+ "loss": 0.8438,
+ "step": 428
+ },
+ {
+ "epoch": 10.81,
+ "learning_rate": 0.000328125,
+ "loss": 0.8862,
+ "step": 429
+ },
+ {
+ "epoch": 10.83,
+ "learning_rate": 0.00032772435897435897,
+ "loss": 0.8631,
+ "step": 430
+ },
+ {
+ "epoch": 10.86,
+ "learning_rate": 0.00032732371794871797,
+ "loss": 0.883,
+ "step": 431
+ },
+ {
+ "epoch": 10.89,
+ "learning_rate": 0.0003269230769230769,
+ "loss": 0.8644,
+ "step": 432
+ },
+ {
+ "epoch": 10.91,
+ "learning_rate": 0.0003265224358974359,
+ "loss": 0.8907,
+ "step": 433
+ },
+ {
+ "epoch": 10.94,
+ "learning_rate": 0.00032612179487179487,
+ "loss": 0.9141,
+ "step": 434
+ },
+ {
+ "epoch": 10.96,
+ "learning_rate": 0.00032572115384615387,
+ "loss": 0.8877,
+ "step": 435
+ },
+ {
+ "epoch": 10.99,
+ "learning_rate": 0.0003253205128205128,
+ "loss": 0.8521,
+ "step": 436
+ },
+ {
+ "epoch": 11.01,
+ "learning_rate": 0.0003249198717948718,
+ "loss": 0.7842,
+ "step": 437
+ },
+ {
+ "epoch": 11.04,
+ "learning_rate": 0.00032451923076923077,
+ "loss": 0.6516,
+ "step": 438
+ },
+ {
+ "epoch": 11.06,
+ "learning_rate": 0.00032411858974358977,
+ "loss": 0.6549,
+ "step": 439
+ },
+ {
+ "epoch": 11.09,
+ "learning_rate": 0.0003237179487179487,
+ "loss": 0.6736,
+ "step": 440
+ },
+ {
+ "epoch": 11.11,
+ "learning_rate": 0.0003233173076923077,
+ "loss": 0.6625,
+ "step": 441
+ },
+ {
+ "epoch": 11.14,
+ "learning_rate": 0.0003229166666666667,
+ "loss": 0.6466,
+ "step": 442
+ },
+ {
+ "epoch": 11.16,
+ "learning_rate": 0.00032251602564102567,
+ "loss": 0.6362,
+ "step": 443
+ },
+ {
+ "epoch": 11.19,
+ "learning_rate": 0.00032211538461538467,
+ "loss": 0.674,
+ "step": 444
+ },
+ {
+ "epoch": 11.21,
+ "learning_rate": 0.0003217147435897436,
+ "loss": 0.6509,
+ "step": 445
+ },
+ {
+ "epoch": 11.24,
+ "learning_rate": 0.0003213141025641026,
+ "loss": 0.618,
+ "step": 446
+ },
+ {
+ "epoch": 11.26,
+ "learning_rate": 0.00032091346153846156,
+ "loss": 0.6654,
+ "step": 447
+ },
+ {
+ "epoch": 11.29,
+ "learning_rate": 0.00032051282051282057,
+ "loss": 0.6632,
+ "step": 448
+ },
+ {
+ "epoch": 11.31,
+ "learning_rate": 0.00032011217948717946,
+ "loss": 0.6599,
+ "step": 449
+ },
+ {
+ "epoch": 11.34,
+ "learning_rate": 0.00031971153846153846,
+ "loss": 0.6537,
+ "step": 450
+ },
+ {
+ "epoch": 11.36,
+ "learning_rate": 0.0003193108974358974,
+ "loss": 0.678,
+ "step": 451
+ },
+ {
+ "epoch": 11.39,
+ "learning_rate": 0.0003189102564102564,
+ "loss": 0.684,
+ "step": 452
+ },
+ {
+ "epoch": 11.41,
+ "learning_rate": 0.00031850961538461536,
+ "loss": 0.6419,
+ "step": 453
+ },
+ {
+ "epoch": 11.44,
+ "learning_rate": 0.00031810897435897436,
+ "loss": 0.6614,
+ "step": 454
+ },
+ {
+ "epoch": 11.46,
+ "learning_rate": 0.0003177083333333333,
+ "loss": 0.7037,
+ "step": 455
+ },
+ {
+ "epoch": 11.49,
+ "learning_rate": 0.0003173076923076923,
+ "loss": 0.6944,
+ "step": 456
+ },
+ {
+ "epoch": 11.51,
+ "learning_rate": 0.00031690705128205126,
+ "loss": 0.6818,
+ "step": 457
+ },
+ {
+ "epoch": 11.54,
+ "learning_rate": 0.00031650641025641026,
+ "loss": 0.692,
+ "step": 458
+ },
+ {
+ "epoch": 11.57,
+ "learning_rate": 0.0003161057692307692,
+ "loss": 0.7028,
+ "step": 459
+ },
+ {
+ "epoch": 11.59,
+ "learning_rate": 0.0003157051282051282,
+ "loss": 0.7009,
+ "step": 460
+ },
+ {
+ "epoch": 11.62,
+ "learning_rate": 0.00031530448717948716,
+ "loss": 0.7325,
+ "step": 461
+ },
+ {
+ "epoch": 11.64,
+ "learning_rate": 0.00031490384615384616,
+ "loss": 0.6738,
+ "step": 462
+ },
+ {
+ "epoch": 11.67,
+ "learning_rate": 0.0003145032051282051,
+ "loss": 0.7008,
+ "step": 463
+ },
+ {
+ "epoch": 11.69,
+ "learning_rate": 0.0003141025641025641,
+ "loss": 0.6838,
+ "step": 464
+ },
+ {
+ "epoch": 11.72,
+ "learning_rate": 0.00031370192307692306,
+ "loss": 0.726,
+ "step": 465
+ },
+ {
+ "epoch": 11.74,
+ "learning_rate": 0.00031330128205128206,
+ "loss": 0.7167,
+ "step": 466
+ },
+ {
+ "epoch": 11.77,
+ "learning_rate": 0.000312900641025641,
+ "loss": 0.7236,
+ "step": 467
+ },
+ {
+ "epoch": 11.79,
+ "learning_rate": 0.0003125,
+ "loss": 0.7076,
+ "step": 468
+ },
+ {
+ "epoch": 11.82,
+ "learning_rate": 0.000312099358974359,
+ "loss": 0.7415,
+ "step": 469
+ },
+ {
+ "epoch": 11.84,
+ "learning_rate": 0.00031169871794871796,
+ "loss": 0.7313,
+ "step": 470
+ },
+ {
+ "epoch": 11.87,
+ "learning_rate": 0.00031129807692307696,
+ "loss": 0.7403,
+ "step": 471
+ },
+ {
+ "epoch": 11.89,
+ "learning_rate": 0.0003108974358974359,
+ "loss": 0.7145,
+ "step": 472
+ },
+ {
+ "epoch": 11.92,
+ "learning_rate": 0.0003104967948717949,
+ "loss": 0.7532,
+ "step": 473
+ },
+ {
+ "epoch": 11.94,
+ "learning_rate": 0.00031009615384615385,
+ "loss": 0.7326,
+ "step": 474
+ },
+ {
+ "epoch": 11.97,
+ "learning_rate": 0.00030969551282051286,
+ "loss": 0.7699,
+ "step": 475
+ },
+ {
+ "epoch": 11.99,
+ "learning_rate": 0.0003092948717948718,
+ "loss": 0.7251,
+ "step": 476
+ },
+ {
+ "epoch": 12.02,
+ "learning_rate": 0.0003088942307692308,
+ "loss": 0.5773,
+ "step": 477
+ },
+ {
+ "epoch": 12.04,
+ "learning_rate": 0.00030849358974358975,
+ "loss": 0.5288,
+ "step": 478
+ },
+ {
+ "epoch": 12.07,
+ "learning_rate": 0.00030809294871794876,
+ "loss": 0.5312,
+ "step": 479
+ },
+ {
+ "epoch": 12.09,
+ "learning_rate": 0.0003076923076923077,
+ "loss": 0.5445,
+ "step": 480
+ },
+ {
+ "epoch": 12.12,
+ "learning_rate": 0.0003072916666666667,
+ "loss": 0.5218,
+ "step": 481
+ },
+ {
+ "epoch": 12.14,
+ "learning_rate": 0.00030689102564102565,
+ "loss": 0.5311,
+ "step": 482
+ },
+ {
+ "epoch": 12.17,
+ "learning_rate": 0.00030649038461538465,
+ "loss": 0.5541,
+ "step": 483
+ },
+ {
+ "epoch": 12.2,
+ "learning_rate": 0.0003060897435897436,
+ "loss": 0.5242,
+ "step": 484
+ },
+ {
+ "epoch": 12.22,
+ "learning_rate": 0.0003056891025641026,
+ "loss": 0.5302,
+ "step": 485
+ },
+ {
+ "epoch": 12.25,
+ "learning_rate": 0.00030528846153846155,
+ "loss": 0.536,
+ "step": 486
+ },
+ {
+ "epoch": 12.27,
+ "learning_rate": 0.00030488782051282055,
+ "loss": 0.5545,
+ "step": 487
+ },
+ {
+ "epoch": 12.3,
+ "learning_rate": 0.00030448717948717945,
+ "loss": 0.5476,
+ "step": 488
+ },
+ {
+ "epoch": 12.32,
+ "learning_rate": 0.00030408653846153845,
+ "loss": 0.5528,
+ "step": 489
+ },
+ {
+ "epoch": 12.35,
+ "learning_rate": 0.0003036858974358974,
+ "loss": 0.5633,
+ "step": 490
+ },
+ {
+ "epoch": 12.37,
+ "learning_rate": 0.0003032852564102564,
+ "loss": 0.571,
+ "step": 491
+ },
+ {
+ "epoch": 12.4,
+ "learning_rate": 0.00030288461538461535,
+ "loss": 0.5637,
+ "step": 492
+ },
+ {
+ "epoch": 12.42,
+ "learning_rate": 0.00030248397435897435,
+ "loss": 0.5613,
+ "step": 493
+ },
+ {
+ "epoch": 12.45,
+ "learning_rate": 0.0003020833333333333,
+ "loss": 0.5999,
+ "step": 494
+ },
+ {
+ "epoch": 12.47,
+ "learning_rate": 0.0003016826923076923,
+ "loss": 0.5633,
+ "step": 495
+ },
+ {
+ "epoch": 12.5,
+ "learning_rate": 0.0003012820512820513,
+ "loss": 0.5811,
+ "step": 496
+ },
+ {
+ "epoch": 12.52,
+ "learning_rate": 0.00030088141025641025,
+ "loss": 0.5711,
+ "step": 497
+ },
+ {
+ "epoch": 12.55,
+ "learning_rate": 0.00030048076923076925,
+ "loss": 0.5669,
+ "step": 498
+ },
+ {
+ "epoch": 12.57,
+ "learning_rate": 0.0003000801282051282,
+ "loss": 0.5424,
+ "step": 499
+ },
+ {
+ "epoch": 12.6,
+ "learning_rate": 0.0002996794871794872,
+ "loss": 0.5592,
+ "step": 500
+ },
+ {
+ "epoch": 12.62,
+ "learning_rate": 0.00029927884615384614,
+ "loss": 0.5645,
+ "step": 501
+ },
+ {
+ "epoch": 12.65,
+ "learning_rate": 0.00029887820512820515,
+ "loss": 0.6089,
+ "step": 502
+ },
+ {
+ "epoch": 12.67,
+ "learning_rate": 0.0002984775641025641,
+ "loss": 0.57,
+ "step": 503
+ },
+ {
+ "epoch": 12.7,
+ "learning_rate": 0.0002980769230769231,
+ "loss": 0.57,
+ "step": 504
+ },
+ {
+ "epoch": 12.72,
+ "learning_rate": 0.00029767628205128204,
+ "loss": 0.5511,
+ "step": 505
+ },
+ {
+ "epoch": 12.75,
+ "learning_rate": 0.00029727564102564105,
+ "loss": 0.5765,
+ "step": 506
+ },
+ {
+ "epoch": 12.77,
+ "learning_rate": 0.000296875,
+ "loss": 0.5896,
+ "step": 507
+ },
+ {
+ "epoch": 12.8,
+ "learning_rate": 0.000296474358974359,
+ "loss": 0.6329,
+ "step": 508
+ },
+ {
+ "epoch": 12.83,
+ "learning_rate": 0.00029607371794871794,
+ "loss": 0.5682,
+ "step": 509
+ },
+ {
+ "epoch": 12.85,
+ "learning_rate": 0.00029567307692307694,
+ "loss": 0.58,
+ "step": 510
+ },
+ {
+ "epoch": 12.88,
+ "learning_rate": 0.0002952724358974359,
+ "loss": 0.6014,
+ "step": 511
+ },
+ {
+ "epoch": 12.9,
+ "learning_rate": 0.0002948717948717949,
+ "loss": 0.6082,
+ "step": 512
+ },
+ {
+ "epoch": 12.93,
+ "learning_rate": 0.00029447115384615384,
+ "loss": 0.6392,
+ "step": 513
+ },
+ {
+ "epoch": 12.95,
+ "learning_rate": 0.00029407051282051284,
+ "loss": 0.609,
+ "step": 514
+ },
+ {
+ "epoch": 12.98,
+ "learning_rate": 0.0002936698717948718,
+ "loss": 0.6247,
+ "step": 515
+ },
+ {
+ "epoch": 13.0,
+ "learning_rate": 0.0002932692307692308,
+ "loss": 0.5943,
+ "step": 516
+ },
+ {
+ "epoch": 13.03,
+ "learning_rate": 0.00029286858974358974,
+ "loss": 0.4499,
+ "step": 517
+ },
+ {
+ "epoch": 13.05,
+ "learning_rate": 0.00029246794871794874,
+ "loss": 0.4412,
+ "step": 518
+ },
+ {
+ "epoch": 13.08,
+ "learning_rate": 0.0002920673076923077,
+ "loss": 0.4409,
+ "step": 519
+ },
+ {
+ "epoch": 13.1,
+ "learning_rate": 0.0002916666666666667,
+ "loss": 0.4505,
+ "step": 520
+ },
+ {
+ "epoch": 13.13,
+ "learning_rate": 0.0002912660256410257,
+ "loss": 0.4489,
+ "step": 521
+ },
+ {
+ "epoch": 13.15,
+ "learning_rate": 0.00029086538461538464,
+ "loss": 0.4235,
+ "step": 522
+ },
+ {
+ "epoch": 13.18,
+ "learning_rate": 0.00029046474358974364,
+ "loss": 0.4276,
+ "step": 523
+ },
+ {
+ "epoch": 13.2,
+ "learning_rate": 0.0002900641025641026,
+ "loss": 0.4544,
+ "step": 524
+ },
+ {
+ "epoch": 13.23,
+ "learning_rate": 0.0002896634615384616,
+ "loss": 0.4625,
+ "step": 525
+ },
+ {
+ "epoch": 13.25,
+ "learning_rate": 0.00028926282051282054,
+ "loss": 0.4266,
+ "step": 526
+ },
+ {
+ "epoch": 13.28,
+ "learning_rate": 0.0002888621794871795,
+ "loss": 0.4459,
+ "step": 527
+ },
+ {
+ "epoch": 13.3,
+ "learning_rate": 0.00028846153846153843,
+ "loss": 0.4261,
+ "step": 528
+ },
+ {
+ "epoch": 13.33,
+ "learning_rate": 0.00028806089743589744,
+ "loss": 0.4376,
+ "step": 529
+ },
+ {
+ "epoch": 13.35,
+ "learning_rate": 0.0002876602564102564,
+ "loss": 0.4418,
+ "step": 530
+ },
+ {
+ "epoch": 13.38,
+ "learning_rate": 0.0002872596153846154,
+ "loss": 0.4667,
+ "step": 531
+ },
+ {
+ "epoch": 13.4,
+ "learning_rate": 0.00028685897435897433,
+ "loss": 0.4504,
+ "step": 532
+ },
+ {
+ "epoch": 13.43,
+ "learning_rate": 0.00028645833333333333,
+ "loss": 0.4691,
+ "step": 533
+ },
+ {
+ "epoch": 13.46,
+ "learning_rate": 0.0002860576923076923,
+ "loss": 0.4684,
+ "step": 534
+ },
+ {
+ "epoch": 13.48,
+ "learning_rate": 0.0002856570512820513,
+ "loss": 0.467,
+ "step": 535
+ },
+ {
+ "epoch": 13.51,
+ "learning_rate": 0.00028525641025641023,
+ "loss": 0.4716,
+ "step": 536
+ },
+ {
+ "epoch": 13.53,
+ "learning_rate": 0.00028485576923076923,
+ "loss": 0.4619,
+ "step": 537
+ },
+ {
+ "epoch": 13.56,
+ "learning_rate": 0.0002844551282051282,
+ "loss": 0.4775,
+ "step": 538
+ },
+ {
+ "epoch": 13.58,
+ "learning_rate": 0.0002840544871794872,
+ "loss": 0.4568,
+ "step": 539
+ },
+ {
+ "epoch": 13.61,
+ "learning_rate": 0.00028365384615384613,
+ "loss": 0.4763,
+ "step": 540
+ },
+ {
+ "epoch": 13.63,
+ "learning_rate": 0.00028325320512820513,
+ "loss": 0.466,
+ "step": 541
+ },
+ {
+ "epoch": 13.66,
+ "learning_rate": 0.0002828525641025641,
+ "loss": 0.4543,
+ "step": 542
+ },
+ {
+ "epoch": 13.68,
+ "learning_rate": 0.0002824519230769231,
+ "loss": 0.4666,
+ "step": 543
+ },
+ {
+ "epoch": 13.71,
+ "learning_rate": 0.00028205128205128203,
+ "loss": 0.489,
+ "step": 544
+ },
+ {
+ "epoch": 13.73,
+ "learning_rate": 0.00028165064102564103,
+ "loss": 0.491,
+ "step": 545
+ },
+ {
+ "epoch": 13.76,
+ "learning_rate": 0.00028125000000000003,
+ "loss": 0.4847,
+ "step": 546
+ },
+ {
+ "epoch": 13.78,
+ "learning_rate": 0.000280849358974359,
+ "loss": 0.4899,
+ "step": 547
+ },
+ {
+ "epoch": 13.81,
+ "learning_rate": 0.000280448717948718,
+ "loss": 0.5169,
+ "step": 548
+ },
+ {
+ "epoch": 13.83,
+ "learning_rate": 0.00028004807692307693,
+ "loss": 0.4968,
+ "step": 549
+ },
+ {
+ "epoch": 13.86,
+ "learning_rate": 0.00027964743589743593,
+ "loss": 0.4774,
+ "step": 550
+ },
+ {
+ "epoch": 13.88,
+ "learning_rate": 0.0002792467948717949,
+ "loss": 0.5105,
+ "step": 551
+ },
+ {
+ "epoch": 13.91,
+ "learning_rate": 0.0002788461538461539,
+ "loss": 0.4997,
+ "step": 552
+ },
+ {
+ "epoch": 13.93,
+ "learning_rate": 0.00027844551282051283,
+ "loss": 0.5088,
+ "step": 553
+ },
+ {
+ "epoch": 13.96,
+ "learning_rate": 0.00027804487179487183,
+ "loss": 0.5091,
+ "step": 554
+ },
+ {
+ "epoch": 13.98,
+ "learning_rate": 0.0002776442307692308,
+ "loss": 0.532,
+ "step": 555
+ },
+ {
+ "epoch": 14.01,
+ "learning_rate": 0.0002772435897435898,
+ "loss": 0.4487,
+ "step": 556
+ },
+ {
+ "epoch": 14.03,
+ "learning_rate": 0.00027684294871794873,
+ "loss": 0.3525,
+ "step": 557
+ },
+ {
+ "epoch": 14.06,
+ "learning_rate": 0.00027644230769230773,
+ "loss": 0.3715,
+ "step": 558
+ },
+ {
+ "epoch": 14.09,
+ "learning_rate": 0.0002760416666666667,
+ "loss": 0.3547,
+ "step": 559
+ },
+ {
+ "epoch": 14.11,
+ "learning_rate": 0.0002756410256410257,
+ "loss": 0.3862,
+ "step": 560
+ },
+ {
+ "epoch": 14.14,
+ "learning_rate": 0.0002752403846153846,
+ "loss": 0.3514,
+ "step": 561
+ },
+ {
+ "epoch": 14.16,
+ "learning_rate": 0.00027483974358974363,
+ "loss": 0.362,
+ "step": 562
+ },
+ {
+ "epoch": 14.19,
+ "learning_rate": 0.0002744391025641026,
+ "loss": 0.3457,
+ "step": 563
+ },
+ {
+ "epoch": 14.21,
+ "learning_rate": 0.0002740384615384616,
+ "loss": 0.3662,
+ "step": 564
+ },
+ {
+ "epoch": 14.24,
+ "learning_rate": 0.0002736378205128205,
+ "loss": 0.3746,
+ "step": 565
+ },
+ {
+ "epoch": 14.26,
+ "learning_rate": 0.00027323717948717947,
+ "loss": 0.3388,
+ "step": 566
+ },
+ {
+ "epoch": 14.29,
+ "learning_rate": 0.0002728365384615384,
+ "loss": 0.3824,
+ "step": 567
+ },
+ {
+ "epoch": 14.31,
+ "learning_rate": 0.0002724358974358974,
+ "loss": 0.3584,
+ "step": 568
+ },
+ {
+ "epoch": 14.34,
+ "learning_rate": 0.00027203525641025637,
+ "loss": 0.3544,
+ "step": 569
+ },
+ {
+ "epoch": 14.36,
+ "learning_rate": 0.00027163461538461537,
+ "loss": 0.3721,
+ "step": 570
+ },
+ {
+ "epoch": 14.39,
+ "learning_rate": 0.0002712339743589743,
+ "loss": 0.3769,
+ "step": 571
+ },
+ {
+ "epoch": 14.41,
+ "learning_rate": 0.0002708333333333333,
+ "loss": 0.3872,
+ "step": 572
+ },
+ {
+ "epoch": 14.44,
+ "learning_rate": 0.0002704326923076923,
+ "loss": 0.363,
+ "step": 573
+ },
+ {
+ "epoch": 14.46,
+ "learning_rate": 0.00027003205128205127,
+ "loss": 0.3842,
+ "step": 574
+ },
+ {
+ "epoch": 14.49,
+ "learning_rate": 0.00026963141025641027,
+ "loss": 0.3737,
+ "step": 575
+ },
+ {
+ "epoch": 14.51,
+ "learning_rate": 0.0002692307692307692,
+ "loss": 0.3981,
+ "step": 576
+ },
+ {
+ "epoch": 14.54,
+ "learning_rate": 0.0002688301282051282,
+ "loss": 0.3656,
+ "step": 577
+ },
+ {
+ "epoch": 14.56,
+ "learning_rate": 0.00026842948717948717,
+ "loss": 0.3852,
+ "step": 578
+ },
+ {
+ "epoch": 14.59,
+ "learning_rate": 0.00026802884615384617,
+ "loss": 0.3865,
+ "step": 579
+ },
+ {
+ "epoch": 14.61,
+ "learning_rate": 0.0002676282051282051,
+ "loss": 0.3747,
+ "step": 580
+ },
+ {
+ "epoch": 14.64,
+ "learning_rate": 0.0002672275641025641,
+ "loss": 0.3729,
+ "step": 581
+ },
+ {
+ "epoch": 14.66,
+ "learning_rate": 0.00026682692307692307,
+ "loss": 0.3785,
+ "step": 582
+ },
+ {
+ "epoch": 14.69,
+ "learning_rate": 0.00026642628205128207,
+ "loss": 0.4018,
+ "step": 583
+ },
+ {
+ "epoch": 14.71,
+ "learning_rate": 0.000266025641025641,
+ "loss": 0.3854,
+ "step": 584
+ },
+ {
+ "epoch": 14.74,
+ "learning_rate": 0.000265625,
+ "loss": 0.4161,
+ "step": 585
+ },
+ {
+ "epoch": 14.77,
+ "learning_rate": 0.00026522435897435897,
+ "loss": 0.4093,
+ "step": 586
+ },
+ {
+ "epoch": 14.79,
+ "learning_rate": 0.00026482371794871797,
+ "loss": 0.4126,
+ "step": 587
+ },
+ {
+ "epoch": 14.82,
+ "learning_rate": 0.0002644230769230769,
+ "loss": 0.3977,
+ "step": 588
+ },
+ {
+ "epoch": 14.84,
+ "learning_rate": 0.0002640224358974359,
+ "loss": 0.4286,
+ "step": 589
+ },
+ {
+ "epoch": 14.87,
+ "learning_rate": 0.00026362179487179487,
+ "loss": 0.4371,
+ "step": 590
+ },
+ {
+ "epoch": 14.89,
+ "learning_rate": 0.00026322115384615387,
+ "loss": 0.4066,
+ "step": 591
+ },
+ {
+ "epoch": 14.92,
+ "learning_rate": 0.0002628205128205128,
+ "loss": 0.433,
+ "step": 592
+ },
+ {
+ "epoch": 14.94,
+ "learning_rate": 0.0002624198717948718,
+ "loss": 0.4361,
+ "step": 593
+ },
+ {
+ "epoch": 14.97,
+ "learning_rate": 0.00026201923076923076,
+ "loss": 0.4276,
+ "step": 594
+ },
+ {
+ "epoch": 14.99,
+ "learning_rate": 0.00026161858974358977,
+ "loss": 0.4401,
+ "step": 595
+ },
+ {
+ "epoch": 15.02,
+ "learning_rate": 0.0002612179487179487,
+ "loss": 0.3514,
+ "step": 596
+ },
+ {
+ "epoch": 15.04,
+ "learning_rate": 0.0002608173076923077,
+ "loss": 0.306,
+ "step": 597
+ },
+ {
+ "epoch": 15.07,
+ "learning_rate": 0.0002604166666666667,
+ "loss": 0.2862,
+ "step": 598
+ },
+ {
+ "epoch": 15.09,
+ "learning_rate": 0.00026001602564102566,
+ "loss": 0.3042,
+ "step": 599
+ },
+ {
+ "epoch": 15.12,
+ "learning_rate": 0.00025961538461538467,
+ "loss": 0.2978,
+ "step": 600
+ },
+ {
+ "epoch": 15.14,
+ "learning_rate": 0.0002592147435897436,
+ "loss": 0.2975,
+ "step": 601
+ },
+ {
+ "epoch": 15.17,
+ "learning_rate": 0.0002588141025641026,
+ "loss": 0.3172,
+ "step": 602
+ },
+ {
+ "epoch": 15.19,
+ "learning_rate": 0.00025841346153846156,
+ "loss": 0.3046,
+ "step": 603
+ },
+ {
+ "epoch": 15.22,
+ "learning_rate": 0.00025801282051282057,
+ "loss": 0.3008,
+ "step": 604
+ },
+ {
+ "epoch": 15.24,
+ "learning_rate": 0.00025761217948717946,
+ "loss": 0.2956,
+ "step": 605
+ },
+ {
+ "epoch": 15.27,
+ "learning_rate": 0.00025721153846153846,
+ "loss": 0.3137,
+ "step": 606
+ },
+ {
+ "epoch": 15.29,
+ "learning_rate": 0.0002568108974358974,
+ "loss": 0.2961,
+ "step": 607
+ },
+ {
+ "epoch": 15.32,
+ "learning_rate": 0.0002564102564102564,
+ "loss": 0.3059,
+ "step": 608
+ },
+ {
+ "epoch": 15.34,
+ "learning_rate": 0.00025600961538461536,
+ "loss": 0.3077,
+ "step": 609
+ },
+ {
+ "epoch": 15.37,
+ "learning_rate": 0.00025560897435897436,
+ "loss": 0.3189,
+ "step": 610
+ },
+ {
+ "epoch": 15.4,
+ "learning_rate": 0.0002552083333333333,
+ "loss": 0.309,
+ "step": 611
+ },
+ {
+ "epoch": 15.42,
+ "learning_rate": 0.0002548076923076923,
+ "loss": 0.3324,
+ "step": 612
+ },
+ {
+ "epoch": 15.45,
+ "learning_rate": 0.00025440705128205126,
+ "loss": 0.3159,
+ "step": 613
+ },
+ {
+ "epoch": 15.47,
+ "learning_rate": 0.00025400641025641026,
+ "loss": 0.3188,
+ "step": 614
+ },
+ {
+ "epoch": 15.5,
+ "learning_rate": 0.0002536057692307692,
+ "loss": 0.3174,
+ "step": 615
+ },
+ {
+ "epoch": 15.52,
+ "learning_rate": 0.0002532051282051282,
+ "loss": 0.3074,
+ "step": 616
+ },
+ {
+ "epoch": 15.55,
+ "learning_rate": 0.00025280448717948716,
+ "loss": 0.3145,
+ "step": 617
+ },
+ {
+ "epoch": 15.57,
+ "learning_rate": 0.00025240384615384616,
+ "loss": 0.317,
+ "step": 618
+ },
+ {
+ "epoch": 15.6,
+ "learning_rate": 0.0002520032051282051,
+ "loss": 0.318,
+ "step": 619
+ },
+ {
+ "epoch": 15.62,
+ "learning_rate": 0.0002516025641025641,
+ "loss": 0.3236,
+ "step": 620
+ },
+ {
+ "epoch": 15.65,
+ "learning_rate": 0.00025120192307692305,
+ "loss": 0.3282,
+ "step": 621
+ },
+ {
+ "epoch": 15.67,
+ "learning_rate": 0.00025080128205128206,
+ "loss": 0.3394,
+ "step": 622
+ },
+ {
+ "epoch": 15.7,
+ "learning_rate": 0.000250400641025641,
+ "loss": 0.3278,
+ "step": 623
+ },
+ {
+ "epoch": 15.72,
+ "learning_rate": 0.00025,
+ "loss": 0.3255,
+ "step": 624
+ },
+ {
+ "epoch": 15.75,
+ "learning_rate": 0.000249599358974359,
+ "loss": 0.3256,
+ "step": 625
+ },
+ {
+ "epoch": 15.77,
+ "learning_rate": 0.00024919871794871795,
+ "loss": 0.3379,
+ "step": 626
+ },
+ {
+ "epoch": 15.8,
+ "learning_rate": 0.00024879807692307696,
+ "loss": 0.3384,
+ "step": 627
+ },
+ {
+ "epoch": 15.82,
+ "learning_rate": 0.0002483974358974359,
+ "loss": 0.3483,
+ "step": 628
+ },
+ {
+ "epoch": 15.85,
+ "learning_rate": 0.0002479967948717949,
+ "loss": 0.3478,
+ "step": 629
+ },
+ {
+ "epoch": 15.87,
+ "learning_rate": 0.00024759615384615385,
+ "loss": 0.337,
+ "step": 630
+ },
+ {
+ "epoch": 15.9,
+ "learning_rate": 0.00024719551282051286,
+ "loss": 0.3287,
+ "step": 631
+ },
+ {
+ "epoch": 15.92,
+ "learning_rate": 0.0002467948717948718,
+ "loss": 0.3478,
+ "step": 632
+ },
+ {
+ "epoch": 15.95,
+ "learning_rate": 0.0002463942307692308,
+ "loss": 0.3567,
+ "step": 633
+ },
+ {
+ "epoch": 15.97,
+ "learning_rate": 0.00024599358974358975,
+ "loss": 0.3359,
+ "step": 634
+ },
+ {
+ "epoch": 16.0,
+ "learning_rate": 0.0002455929487179487,
+ "loss": 0.3492,
+ "step": 635
+ },
+ {
+ "epoch": 16.03,
+ "learning_rate": 0.0002451923076923077,
+ "loss": 0.2497,
+ "step": 636
+ },
+ {
+ "epoch": 16.05,
+ "learning_rate": 0.00024479166666666665,
+ "loss": 0.2416,
+ "step": 637
+ },
+ {
+ "epoch": 16.08,
+ "learning_rate": 0.00024439102564102565,
+ "loss": 0.2576,
+ "step": 638
+ },
+ {
+ "epoch": 16.1,
+ "learning_rate": 0.00024399038461538463,
+ "loss": 0.243,
+ "step": 639
+ },
+ {
+ "epoch": 16.13,
+ "learning_rate": 0.0002435897435897436,
+ "loss": 0.2527,
+ "step": 640
+ },
+ {
+ "epoch": 16.15,
+ "learning_rate": 0.00024318910256410258,
+ "loss": 0.2606,
+ "step": 641
+ },
+ {
+ "epoch": 16.18,
+ "learning_rate": 0.00024278846153846155,
+ "loss": 0.2555,
+ "step": 642
+ },
+ {
+ "epoch": 16.2,
+ "learning_rate": 0.00024238782051282052,
+ "loss": 0.2581,
+ "step": 643
+ },
+ {
+ "epoch": 16.23,
+ "learning_rate": 0.0002419871794871795,
+ "loss": 0.2508,
+ "step": 644
+ },
+ {
+ "epoch": 16.25,
+ "learning_rate": 0.00024158653846153847,
+ "loss": 0.281,
+ "step": 645
+ },
+ {
+ "epoch": 16.28,
+ "learning_rate": 0.00024118589743589745,
+ "loss": 0.2626,
+ "step": 646
+ },
+ {
+ "epoch": 16.3,
+ "learning_rate": 0.00024078525641025642,
+ "loss": 0.2668,
+ "step": 647
+ },
+ {
+ "epoch": 16.33,
+ "learning_rate": 0.0002403846153846154,
+ "loss": 0.2675,
+ "step": 648
+ },
+ {
+ "epoch": 16.35,
+ "learning_rate": 0.00023998397435897437,
+ "loss": 0.2491,
+ "step": 649
+ },
+ {
+ "epoch": 16.38,
+ "learning_rate": 0.00023958333333333335,
+ "loss": 0.2459,
+ "step": 650
+ },
+ {
+ "epoch": 16.4,
+ "learning_rate": 0.00023918269230769232,
+ "loss": 0.2515,
+ "step": 651
+ },
+ {
+ "epoch": 16.43,
+ "learning_rate": 0.0002387820512820513,
+ "loss": 0.2612,
+ "step": 652
+ },
+ {
+ "epoch": 16.45,
+ "learning_rate": 0.00023838141025641027,
+ "loss": 0.2612,
+ "step": 653
+ },
+ {
+ "epoch": 16.48,
+ "learning_rate": 0.00023798076923076922,
+ "loss": 0.265,
+ "step": 654
+ },
+ {
+ "epoch": 16.5,
+ "learning_rate": 0.0002375801282051282,
+ "loss": 0.2634,
+ "step": 655
+ },
+ {
+ "epoch": 16.53,
+ "learning_rate": 0.00023717948717948717,
+ "loss": 0.2839,
+ "step": 656
+ },
+ {
+ "epoch": 16.55,
+ "learning_rate": 0.00023677884615384614,
+ "loss": 0.2595,
+ "step": 657
+ },
+ {
+ "epoch": 16.58,
+ "learning_rate": 0.00023637820512820512,
+ "loss": 0.2764,
+ "step": 658
+ },
+ {
+ "epoch": 16.6,
+ "learning_rate": 0.0002359775641025641,
+ "loss": 0.2553,
+ "step": 659
+ },
+ {
+ "epoch": 16.63,
+ "learning_rate": 0.00023557692307692307,
+ "loss": 0.2716,
+ "step": 660
+ },
+ {
+ "epoch": 16.66,
+ "learning_rate": 0.00023517628205128204,
+ "loss": 0.289,
+ "step": 661
+ },
+ {
+ "epoch": 16.68,
+ "learning_rate": 0.00023477564102564102,
+ "loss": 0.2483,
+ "step": 662
+ },
+ {
+ "epoch": 16.71,
+ "learning_rate": 0.000234375,
+ "loss": 0.269,
+ "step": 663
+ },
+ {
+ "epoch": 16.73,
+ "learning_rate": 0.000233974358974359,
+ "loss": 0.2738,
+ "step": 664
+ },
+ {
+ "epoch": 16.76,
+ "learning_rate": 0.00023357371794871797,
+ "loss": 0.2646,
+ "step": 665
+ },
+ {
+ "epoch": 16.78,
+ "learning_rate": 0.00023317307692307694,
+ "loss": 0.2793,
+ "step": 666
+ },
+ {
+ "epoch": 16.81,
+ "learning_rate": 0.00023277243589743592,
+ "loss": 0.2666,
+ "step": 667
+ },
+ {
+ "epoch": 16.83,
+ "learning_rate": 0.0002323717948717949,
+ "loss": 0.2799,
+ "step": 668
+ },
+ {
+ "epoch": 16.86,
+ "learning_rate": 0.00023197115384615387,
+ "loss": 0.2677,
+ "step": 669
+ },
+ {
+ "epoch": 16.88,
+ "learning_rate": 0.00023157051282051284,
+ "loss": 0.2889,
+ "step": 670
+ },
+ {
+ "epoch": 16.91,
+ "learning_rate": 0.00023116987179487182,
+ "loss": 0.2869,
+ "step": 671
+ },
+ {
+ "epoch": 16.93,
+ "learning_rate": 0.0002307692307692308,
+ "loss": 0.2814,
+ "step": 672
+ },
+ {
+ "epoch": 16.96,
+ "learning_rate": 0.00023036858974358974,
+ "loss": 0.2881,
+ "step": 673
+ },
+ {
+ "epoch": 16.98,
+ "learning_rate": 0.0002299679487179487,
+ "loss": 0.2777,
+ "step": 674
+ },
+ {
+ "epoch": 17.01,
+ "learning_rate": 0.0002295673076923077,
+ "loss": 0.2713,
+ "step": 675
+ },
+ {
+ "epoch": 17.03,
+ "learning_rate": 0.00022916666666666666,
+ "loss": 0.2176,
+ "step": 676
+ },
+ {
+ "epoch": 17.06,
+ "learning_rate": 0.00022876602564102564,
+ "loss": 0.2032,
+ "step": 677
+ },
+ {
+ "epoch": 17.08,
+ "learning_rate": 0.0002283653846153846,
+ "loss": 0.2207,
+ "step": 678
+ },
+ {
+ "epoch": 17.11,
+ "learning_rate": 0.0002279647435897436,
+ "loss": 0.2183,
+ "step": 679
+ },
+ {
+ "epoch": 17.13,
+ "learning_rate": 0.00022756410256410256,
+ "loss": 0.2238,
+ "step": 680
+ },
+ {
+ "epoch": 17.16,
+ "learning_rate": 0.00022716346153846154,
+ "loss": 0.2126,
+ "step": 681
+ },
+ {
+ "epoch": 17.18,
+ "learning_rate": 0.0002267628205128205,
+ "loss": 0.2104,
+ "step": 682
+ },
+ {
+ "epoch": 17.21,
+ "learning_rate": 0.00022636217948717949,
+ "loss": 0.2151,
+ "step": 683
+ },
+ {
+ "epoch": 17.23,
+ "learning_rate": 0.00022596153846153846,
+ "loss": 0.2093,
+ "step": 684
+ },
+ {
+ "epoch": 17.26,
+ "learning_rate": 0.00022556089743589743,
+ "loss": 0.2308,
+ "step": 685
+ },
+ {
+ "epoch": 17.29,
+ "learning_rate": 0.0002251602564102564,
+ "loss": 0.2276,
+ "step": 686
+ },
+ {
+ "epoch": 17.31,
+ "learning_rate": 0.00022475961538461538,
+ "loss": 0.2328,
+ "step": 687
+ },
+ {
+ "epoch": 17.34,
+ "learning_rate": 0.00022435897435897436,
+ "loss": 0.2076,
+ "step": 688
+ },
+ {
+ "epoch": 17.36,
+ "learning_rate": 0.00022395833333333336,
+ "loss": 0.2125,
+ "step": 689
+ },
+ {
+ "epoch": 17.39,
+ "learning_rate": 0.00022355769230769234,
+ "loss": 0.2181,
+ "step": 690
+ },
+ {
+ "epoch": 17.41,
+ "learning_rate": 0.0002231570512820513,
+ "loss": 0.2153,
+ "step": 691
+ },
+ {
+ "epoch": 17.44,
+ "learning_rate": 0.00022275641025641028,
+ "loss": 0.2122,
+ "step": 692
+ },
+ {
+ "epoch": 17.46,
+ "learning_rate": 0.00022235576923076923,
+ "loss": 0.2262,
+ "step": 693
+ },
+ {
+ "epoch": 17.49,
+ "learning_rate": 0.0002219551282051282,
+ "loss": 0.2206,
+ "step": 694
+ },
+ {
+ "epoch": 17.51,
+ "learning_rate": 0.00022155448717948718,
+ "loss": 0.2223,
+ "step": 695
+ },
+ {
+ "epoch": 17.54,
+ "learning_rate": 0.00022115384615384616,
+ "loss": 0.2306,
+ "step": 696
+ },
+ {
+ "epoch": 17.56,
+ "learning_rate": 0.00022075320512820513,
+ "loss": 0.213,
+ "step": 697
+ },
+ {
+ "epoch": 17.59,
+ "learning_rate": 0.0002203525641025641,
+ "loss": 0.2198,
+ "step": 698
+ },
+ {
+ "epoch": 17.61,
+ "learning_rate": 0.00021995192307692308,
+ "loss": 0.2299,
+ "step": 699
+ },
+ {
+ "epoch": 17.64,
+ "learning_rate": 0.00021955128205128206,
+ "loss": 0.2213,
+ "step": 700
+ },
+ {
+ "epoch": 17.66,
+ "learning_rate": 0.00021915064102564103,
+ "loss": 0.2183,
+ "step": 701
+ },
+ {
+ "epoch": 17.69,
+ "learning_rate": 0.00021875,
+ "loss": 0.2301,
+ "step": 702
+ },
+ {
+ "epoch": 17.71,
+ "learning_rate": 0.00021834935897435898,
+ "loss": 0.2348,
+ "step": 703
+ },
+ {
+ "epoch": 17.74,
+ "learning_rate": 0.00021794871794871795,
+ "loss": 0.229,
+ "step": 704
+ },
+ {
+ "epoch": 17.76,
+ "learning_rate": 0.00021754807692307693,
+ "loss": 0.2397,
+ "step": 705
+ },
+ {
+ "epoch": 17.79,
+ "learning_rate": 0.0002171474358974359,
+ "loss": 0.2254,
+ "step": 706
+ },
+ {
+ "epoch": 17.81,
+ "learning_rate": 0.00021674679487179488,
+ "loss": 0.2306,
+ "step": 707
+ },
+ {
+ "epoch": 17.84,
+ "learning_rate": 0.00021634615384615385,
+ "loss": 0.2388,
+ "step": 708
+ },
+ {
+ "epoch": 17.86,
+ "learning_rate": 0.00021594551282051283,
+ "loss": 0.2374,
+ "step": 709
+ },
+ {
+ "epoch": 17.89,
+ "learning_rate": 0.0002155448717948718,
+ "loss": 0.2412,
+ "step": 710
+ },
+ {
+ "epoch": 17.91,
+ "learning_rate": 0.00021514423076923078,
+ "loss": 0.2203,
+ "step": 711
+ },
+ {
+ "epoch": 17.94,
+ "learning_rate": 0.00021474358974358972,
+ "loss": 0.2359,
+ "step": 712
+ },
+ {
+ "epoch": 17.97,
+ "learning_rate": 0.0002143429487179487,
+ "loss": 0.2343,
+ "step": 713
+ },
+ {
+ "epoch": 17.99,
+ "learning_rate": 0.00021394230769230767,
+ "loss": 0.2397,
+ "step": 714
+ },
+ {
+ "epoch": 18.02,
+ "learning_rate": 0.00021354166666666665,
+ "loss": 0.2055,
+ "step": 715
+ },
+ {
+ "epoch": 18.04,
+ "learning_rate": 0.00021314102564102565,
+ "loss": 0.1723,
+ "step": 716
+ },
+ {
+ "epoch": 18.07,
+ "learning_rate": 0.00021274038461538463,
+ "loss": 0.1782,
+ "step": 717
+ },
+ {
+ "epoch": 18.09,
+ "learning_rate": 0.0002123397435897436,
+ "loss": 0.1813,
+ "step": 718
+ },
+ {
+ "epoch": 18.12,
+ "learning_rate": 0.00021193910256410257,
+ "loss": 0.1834,
+ "step": 719
+ },
+ {
+ "epoch": 18.14,
+ "learning_rate": 0.00021153846153846155,
+ "loss": 0.1759,
+ "step": 720
+ },
+ {
+ "epoch": 18.17,
+ "learning_rate": 0.00021113782051282052,
+ "loss": 0.1855,
+ "step": 721
+ },
+ {
+ "epoch": 18.19,
+ "learning_rate": 0.0002107371794871795,
+ "loss": 0.1756,
+ "step": 722
+ },
+ {
+ "epoch": 18.22,
+ "learning_rate": 0.00021033653846153847,
+ "loss": 0.1783,
+ "step": 723
+ },
+ {
+ "epoch": 18.24,
+ "learning_rate": 0.00020993589743589745,
+ "loss": 0.1899,
+ "step": 724
+ },
+ {
+ "epoch": 18.27,
+ "learning_rate": 0.00020953525641025642,
+ "loss": 0.1817,
+ "step": 725
+ },
+ {
+ "epoch": 18.29,
+ "learning_rate": 0.0002091346153846154,
+ "loss": 0.1856,
+ "step": 726
+ },
+ {
+ "epoch": 18.32,
+ "learning_rate": 0.00020873397435897437,
+ "loss": 0.1849,
+ "step": 727
+ },
+ {
+ "epoch": 18.34,
+ "learning_rate": 0.00020833333333333335,
+ "loss": 0.1758,
+ "step": 728
+ },
+ {
+ "epoch": 18.37,
+ "learning_rate": 0.00020793269230769232,
+ "loss": 0.1907,
+ "step": 729
+ },
+ {
+ "epoch": 18.39,
+ "learning_rate": 0.0002075320512820513,
+ "loss": 0.1836,
+ "step": 730
+ },
+ {
+ "epoch": 18.42,
+ "learning_rate": 0.00020713141025641027,
+ "loss": 0.1828,
+ "step": 731
+ },
+ {
+ "epoch": 18.44,
+ "learning_rate": 0.00020673076923076922,
+ "loss": 0.1916,
+ "step": 732
+ },
+ {
+ "epoch": 18.47,
+ "learning_rate": 0.0002063301282051282,
+ "loss": 0.1849,
+ "step": 733
+ },
+ {
+ "epoch": 18.49,
+ "learning_rate": 0.00020592948717948717,
+ "loss": 0.1929,
+ "step": 734
+ },
+ {
+ "epoch": 18.52,
+ "learning_rate": 0.00020552884615384614,
+ "loss": 0.1779,
+ "step": 735
+ },
+ {
+ "epoch": 18.54,
+ "learning_rate": 0.00020512820512820512,
+ "loss": 0.1805,
+ "step": 736
+ },
+ {
+ "epoch": 18.57,
+ "learning_rate": 0.0002047275641025641,
+ "loss": 0.1891,
+ "step": 737
+ },
+ {
+ "epoch": 18.6,
+ "learning_rate": 0.00020432692307692307,
+ "loss": 0.1932,
+ "step": 738
+ },
+ {
+ "epoch": 18.62,
+ "learning_rate": 0.00020392628205128204,
+ "loss": 0.1923,
+ "step": 739
+ },
+ {
+ "epoch": 18.65,
+ "learning_rate": 0.00020352564102564102,
+ "loss": 0.1882,
+ "step": 740
+ },
+ {
+ "epoch": 18.67,
+ "learning_rate": 0.00020312500000000002,
+ "loss": 0.2012,
+ "step": 741
+ },
+ {
+ "epoch": 18.7,
+ "learning_rate": 0.000202724358974359,
+ "loss": 0.1963,
+ "step": 742
+ },
+ {
+ "epoch": 18.72,
+ "learning_rate": 0.00020232371794871797,
+ "loss": 0.202,
+ "step": 743
+ },
+ {
+ "epoch": 18.75,
+ "learning_rate": 0.00020192307692307694,
+ "loss": 0.1996,
+ "step": 744
+ },
+ {
+ "epoch": 18.77,
+ "learning_rate": 0.00020152243589743592,
+ "loss": 0.2056,
+ "step": 745
+ },
+ {
+ "epoch": 18.8,
+ "learning_rate": 0.0002011217948717949,
+ "loss": 0.1946,
+ "step": 746
+ },
+ {
+ "epoch": 18.82,
+ "learning_rate": 0.00020072115384615387,
+ "loss": 0.1954,
+ "step": 747
+ },
+ {
+ "epoch": 18.85,
+ "learning_rate": 0.00020032051282051284,
+ "loss": 0.1966,
+ "step": 748
+ },
+ {
+ "epoch": 18.87,
+ "learning_rate": 0.00019991987179487182,
+ "loss": 0.2077,
+ "step": 749
+ },
+ {
+ "epoch": 18.9,
+ "learning_rate": 0.0001995192307692308,
+ "loss": 0.2,
+ "step": 750
+ },
+ {
+ "epoch": 18.92,
+ "learning_rate": 0.00019911858974358974,
+ "loss": 0.2092,
+ "step": 751
+ },
+ {
+ "epoch": 18.95,
+ "learning_rate": 0.0001987179487179487,
+ "loss": 0.2023,
+ "step": 752
+ },
+ {
+ "epoch": 18.97,
+ "learning_rate": 0.0001983173076923077,
+ "loss": 0.2056,
+ "step": 753
+ },
+ {
+ "epoch": 19.0,
+ "learning_rate": 0.00019791666666666666,
+ "loss": 0.2059,
+ "step": 754
+ },
+ {
+ "epoch": 19.02,
+ "learning_rate": 0.00019751602564102564,
+ "loss": 0.1581,
+ "step": 755
+ },
+ {
+ "epoch": 19.05,
+ "learning_rate": 0.0001971153846153846,
+ "loss": 0.155,
+ "step": 756
+ },
+ {
+ "epoch": 19.07,
+ "learning_rate": 0.00019671474358974359,
+ "loss": 0.156,
+ "step": 757
+ },
+ {
+ "epoch": 19.1,
+ "learning_rate": 0.00019631410256410256,
+ "loss": 0.161,
+ "step": 758
+ },
+ {
+ "epoch": 19.12,
+ "learning_rate": 0.00019591346153846154,
+ "loss": 0.1608,
+ "step": 759
+ },
+ {
+ "epoch": 19.15,
+ "learning_rate": 0.0001955128205128205,
+ "loss": 0.1576,
+ "step": 760
+ },
+ {
+ "epoch": 19.17,
+ "learning_rate": 0.00019511217948717948,
+ "loss": 0.1539,
+ "step": 761
+ },
+ {
+ "epoch": 19.2,
+ "learning_rate": 0.00019471153846153846,
+ "loss": 0.1549,
+ "step": 762
+ },
+ {
+ "epoch": 19.23,
+ "learning_rate": 0.00019431089743589743,
+ "loss": 0.1593,
+ "step": 763
+ },
+ {
+ "epoch": 19.25,
+ "learning_rate": 0.0001939102564102564,
+ "loss": 0.162,
+ "step": 764
+ },
+ {
+ "epoch": 19.28,
+ "learning_rate": 0.00019350961538461538,
+ "loss": 0.1591,
+ "step": 765
+ },
+ {
+ "epoch": 19.3,
+ "learning_rate": 0.00019310897435897436,
+ "loss": 0.1569,
+ "step": 766
+ },
+ {
+ "epoch": 19.33,
+ "learning_rate": 0.00019270833333333336,
+ "loss": 0.1628,
+ "step": 767
+ },
+ {
+ "epoch": 19.35,
+ "learning_rate": 0.00019230769230769233,
+ "loss": 0.1529,
+ "step": 768
+ },
+ {
+ "epoch": 19.38,
+ "learning_rate": 0.0001919070512820513,
+ "loss": 0.1606,
+ "step": 769
+ },
+ {
+ "epoch": 19.4,
+ "learning_rate": 0.00019150641025641028,
+ "loss": 0.1564,
+ "step": 770
+ },
+ {
+ "epoch": 19.43,
+ "learning_rate": 0.00019110576923076923,
+ "loss": 0.1599,
+ "step": 771
+ },
+ {
+ "epoch": 19.45,
+ "learning_rate": 0.0001907051282051282,
+ "loss": 0.1653,
+ "step": 772
+ },
+ {
+ "epoch": 19.48,
+ "learning_rate": 0.00019030448717948718,
+ "loss": 0.1587,
+ "step": 773
+ },
+ {
+ "epoch": 19.5,
+ "learning_rate": 0.00018990384615384616,
+ "loss": 0.1683,
+ "step": 774
+ },
+ {
+ "epoch": 19.53,
+ "learning_rate": 0.00018950320512820513,
+ "loss": 0.157,
+ "step": 775
+ },
+ {
+ "epoch": 19.55,
+ "learning_rate": 0.0001891025641025641,
+ "loss": 0.1588,
+ "step": 776
+ },
+ {
+ "epoch": 19.58,
+ "learning_rate": 0.00018870192307692308,
+ "loss": 0.1709,
+ "step": 777
+ },
+ {
+ "epoch": 19.6,
+ "learning_rate": 0.00018830128205128205,
+ "loss": 0.1635,
+ "step": 778
+ },
+ {
+ "epoch": 19.63,
+ "learning_rate": 0.00018790064102564103,
+ "loss": 0.1581,
+ "step": 779
+ },
+ {
+ "epoch": 19.65,
+ "learning_rate": 0.0001875,
+ "loss": 0.1664,
+ "step": 780
+ },
+ {
+ "epoch": 19.68,
+ "learning_rate": 0.00018709935897435898,
+ "loss": 0.1712,
+ "step": 781
+ },
+ {
+ "epoch": 19.7,
+ "learning_rate": 0.00018669871794871795,
+ "loss": 0.165,
+ "step": 782
+ },
+ {
+ "epoch": 19.73,
+ "learning_rate": 0.00018629807692307693,
+ "loss": 0.1692,
+ "step": 783
+ },
+ {
+ "epoch": 19.75,
+ "learning_rate": 0.0001858974358974359,
+ "loss": 0.1658,
+ "step": 784
+ },
+ {
+ "epoch": 19.78,
+ "learning_rate": 0.00018549679487179488,
+ "loss": 0.1702,
+ "step": 785
+ },
+ {
+ "epoch": 19.8,
+ "learning_rate": 0.00018509615384615385,
+ "loss": 0.1672,
+ "step": 786
+ },
+ {
+ "epoch": 19.83,
+ "learning_rate": 0.00018469551282051283,
+ "loss": 0.1708,
+ "step": 787
+ },
+ {
+ "epoch": 19.86,
+ "learning_rate": 0.0001842948717948718,
+ "loss": 0.1705,
+ "step": 788
+ },
+ {
+ "epoch": 19.88,
+ "learning_rate": 0.00018389423076923078,
+ "loss": 0.1853,
+ "step": 789
+ },
+ {
+ "epoch": 19.91,
+ "learning_rate": 0.00018349358974358972,
+ "loss": 0.1701,
+ "step": 790
+ },
+ {
+ "epoch": 19.93,
+ "learning_rate": 0.0001830929487179487,
+ "loss": 0.1726,
+ "step": 791
+ },
+ {
+ "epoch": 19.96,
+ "learning_rate": 0.00018269230769230767,
+ "loss": 0.1829,
+ "step": 792
+ },
+ {
+ "epoch": 19.98,
+ "learning_rate": 0.00018229166666666665,
+ "loss": 0.1749,
+ "step": 793
+ },
+ {
+ "epoch": 20.01,
+ "learning_rate": 0.00018189102564102565,
+ "loss": 0.1559,
+ "step": 794
+ },
+ {
+ "epoch": 20.03,
+ "learning_rate": 0.00018149038461538462,
+ "loss": 0.1424,
+ "step": 795
+ },
+ {
+ "epoch": 20.06,
+ "learning_rate": 0.0001810897435897436,
+ "loss": 0.1386,
+ "step": 796
+ },
+ {
+ "epoch": 20.08,
+ "learning_rate": 0.00018068910256410257,
+ "loss": 0.1388,
+ "step": 797
+ },
+ {
+ "epoch": 20.11,
+ "learning_rate": 0.00018028846153846155,
+ "loss": 0.1385,
+ "step": 798
+ },
+ {
+ "epoch": 20.13,
+ "learning_rate": 0.00017988782051282052,
+ "loss": 0.1349,
+ "step": 799
+ },
+ {
+ "epoch": 20.16,
+ "learning_rate": 0.0001794871794871795,
+ "loss": 0.1357,
+ "step": 800
+ },
+ {
+ "epoch": 20.18,
+ "learning_rate": 0.00017908653846153847,
+ "loss": 0.1384,
+ "step": 801
+ },
+ {
+ "epoch": 20.21,
+ "learning_rate": 0.00017868589743589745,
+ "loss": 0.1425,
+ "step": 802
+ },
+ {
+ "epoch": 20.23,
+ "learning_rate": 0.00017828525641025642,
+ "loss": 0.1337,
+ "step": 803
+ },
+ {
+ "epoch": 20.26,
+ "learning_rate": 0.0001778846153846154,
+ "loss": 0.136,
+ "step": 804
+ },
+ {
+ "epoch": 20.28,
+ "learning_rate": 0.00017748397435897437,
+ "loss": 0.1382,
+ "step": 805
+ },
+ {
+ "epoch": 20.31,
+ "learning_rate": 0.00017708333333333335,
+ "loss": 0.1396,
+ "step": 806
+ },
+ {
+ "epoch": 20.33,
+ "learning_rate": 0.00017668269230769232,
+ "loss": 0.1411,
+ "step": 807
+ },
+ {
+ "epoch": 20.36,
+ "learning_rate": 0.0001762820512820513,
+ "loss": 0.1375,
+ "step": 808
+ },
+ {
+ "epoch": 20.38,
+ "learning_rate": 0.00017588141025641027,
+ "loss": 0.1457,
+ "step": 809
+ },
+ {
+ "epoch": 20.41,
+ "learning_rate": 0.00017548076923076922,
+ "loss": 0.1447,
+ "step": 810
+ },
+ {
+ "epoch": 20.43,
+ "learning_rate": 0.0001750801282051282,
+ "loss": 0.1381,
+ "step": 811
+ },
+ {
+ "epoch": 20.46,
+ "learning_rate": 0.00017467948717948717,
+ "loss": 0.1443,
+ "step": 812
+ },
+ {
+ "epoch": 20.49,
+ "learning_rate": 0.00017427884615384614,
+ "loss": 0.1427,
+ "step": 813
+ },
+ {
+ "epoch": 20.51,
+ "learning_rate": 0.00017387820512820512,
+ "loss": 0.1425,
+ "step": 814
+ },
+ {
+ "epoch": 20.54,
+ "learning_rate": 0.0001734775641025641,
+ "loss": 0.1428,
+ "step": 815
+ },
+ {
+ "epoch": 20.56,
+ "learning_rate": 0.00017307692307692307,
+ "loss": 0.1454,
+ "step": 816
+ },
+ {
+ "epoch": 20.59,
+ "learning_rate": 0.00017267628205128204,
+ "loss": 0.1422,
+ "step": 817
+ },
+ {
+ "epoch": 20.61,
+ "learning_rate": 0.00017227564102564102,
+ "loss": 0.1426,
+ "step": 818
+ },
+ {
+ "epoch": 20.64,
+ "learning_rate": 0.000171875,
+ "loss": 0.1453,
+ "step": 819
+ },
+ {
+ "epoch": 20.66,
+ "learning_rate": 0.000171474358974359,
+ "loss": 0.1472,
+ "step": 820
+ },
+ {
+ "epoch": 20.69,
+ "learning_rate": 0.00017107371794871797,
+ "loss": 0.1479,
+ "step": 821
+ },
+ {
+ "epoch": 20.71,
+ "learning_rate": 0.00017067307692307694,
+ "loss": 0.1387,
+ "step": 822
+ },
+ {
+ "epoch": 20.74,
+ "learning_rate": 0.00017027243589743592,
+ "loss": 0.15,
+ "step": 823
+ },
+ {
+ "epoch": 20.76,
+ "learning_rate": 0.0001698717948717949,
+ "loss": 0.1429,
+ "step": 824
+ },
+ {
+ "epoch": 20.79,
+ "learning_rate": 0.00016947115384615387,
+ "loss": 0.1532,
+ "step": 825
+ },
+ {
+ "epoch": 20.81,
+ "learning_rate": 0.00016907051282051284,
+ "loss": 0.1452,
+ "step": 826
+ },
+ {
+ "epoch": 20.84,
+ "learning_rate": 0.00016866987179487181,
+ "loss": 0.1445,
+ "step": 827
+ },
+ {
+ "epoch": 20.86,
+ "learning_rate": 0.0001682692307692308,
+ "loss": 0.1461,
+ "step": 828
+ },
+ {
+ "epoch": 20.89,
+ "learning_rate": 0.00016786858974358974,
+ "loss": 0.1544,
+ "step": 829
+ },
+ {
+ "epoch": 20.91,
+ "learning_rate": 0.0001674679487179487,
+ "loss": 0.1476,
+ "step": 830
+ },
+ {
+ "epoch": 20.94,
+ "learning_rate": 0.00016706730769230769,
+ "loss": 0.1535,
+ "step": 831
+ },
+ {
+ "epoch": 20.96,
+ "learning_rate": 0.00016666666666666666,
+ "loss": 0.1498,
+ "step": 832
+ },
+ {
+ "epoch": 20.99,
+ "learning_rate": 0.00016626602564102564,
+ "loss": 0.1514,
+ "step": 833
+ },
+ {
+ "epoch": 21.01,
+ "learning_rate": 0.0001658653846153846,
+ "loss": 0.1405,
+ "step": 834
+ },
+ {
+ "epoch": 21.04,
+ "learning_rate": 0.00016546474358974359,
+ "loss": 0.1218,
+ "step": 835
+ },
+ {
+ "epoch": 21.06,
+ "learning_rate": 0.00016506410256410256,
+ "loss": 0.1228,
+ "step": 836
+ },
+ {
+ "epoch": 21.09,
+ "learning_rate": 0.00016466346153846153,
+ "loss": 0.119,
+ "step": 837
+ },
+ {
+ "epoch": 21.11,
+ "learning_rate": 0.0001642628205128205,
+ "loss": 0.1278,
+ "step": 838
+ },
+ {
+ "epoch": 21.14,
+ "learning_rate": 0.00016386217948717948,
+ "loss": 0.1205,
+ "step": 839
+ },
+ {
+ "epoch": 21.17,
+ "learning_rate": 0.00016346153846153846,
+ "loss": 0.1234,
+ "step": 840
+ },
+ {
+ "epoch": 21.19,
+ "learning_rate": 0.00016306089743589743,
+ "loss": 0.1226,
+ "step": 841
+ },
+ {
+ "epoch": 21.22,
+ "learning_rate": 0.0001626602564102564,
+ "loss": 0.1216,
+ "step": 842
+ },
+ {
+ "epoch": 21.24,
+ "learning_rate": 0.00016225961538461538,
+ "loss": 0.1221,
+ "step": 843
+ },
+ {
+ "epoch": 21.27,
+ "learning_rate": 0.00016185897435897436,
+ "loss": 0.1234,
+ "step": 844
+ },
+ {
+ "epoch": 21.29,
+ "learning_rate": 0.00016145833333333336,
+ "loss": 0.1226,
+ "step": 845
+ },
+ {
+ "epoch": 21.32,
+ "learning_rate": 0.00016105769230769233,
+ "loss": 0.1224,
+ "step": 846
+ },
+ {
+ "epoch": 21.34,
+ "learning_rate": 0.0001606570512820513,
+ "loss": 0.1266,
+ "step": 847
+ },
+ {
+ "epoch": 21.37,
+ "learning_rate": 0.00016025641025641028,
+ "loss": 0.1267,
+ "step": 848
+ },
+ {
+ "epoch": 21.39,
+ "learning_rate": 0.00015985576923076923,
+ "loss": 0.129,
+ "step": 849
+ },
+ {
+ "epoch": 21.42,
+ "learning_rate": 0.0001594551282051282,
+ "loss": 0.1213,
+ "step": 850
+ },
+ {
+ "epoch": 21.44,
+ "learning_rate": 0.00015905448717948718,
+ "loss": 0.1273,
+ "step": 851
+ },
+ {
+ "epoch": 21.47,
+ "learning_rate": 0.00015865384615384616,
+ "loss": 0.1352,
+ "step": 852
+ },
+ {
+ "epoch": 21.49,
+ "learning_rate": 0.00015825320512820513,
+ "loss": 0.1328,
+ "step": 853
+ },
+ {
+ "epoch": 21.52,
+ "learning_rate": 0.0001578525641025641,
+ "loss": 0.1253,
+ "step": 854
+ },
+ {
+ "epoch": 21.54,
+ "learning_rate": 0.00015745192307692308,
+ "loss": 0.1267,
+ "step": 855
+ },
+ {
+ "epoch": 21.57,
+ "learning_rate": 0.00015705128205128205,
+ "loss": 0.1325,
+ "step": 856
+ },
+ {
+ "epoch": 21.59,
+ "learning_rate": 0.00015665064102564103,
+ "loss": 0.1267,
+ "step": 857
+ },
+ {
+ "epoch": 21.62,
+ "learning_rate": 0.00015625,
+ "loss": 0.1326,
+ "step": 858
+ },
+ {
+ "epoch": 21.64,
+ "learning_rate": 0.00015584935897435898,
+ "loss": 0.1319,
+ "step": 859
+ },
+ {
+ "epoch": 21.67,
+ "learning_rate": 0.00015544871794871795,
+ "loss": 0.1262,
+ "step": 860
+ },
+ {
+ "epoch": 21.69,
+ "learning_rate": 0.00015504807692307693,
+ "loss": 0.1334,
+ "step": 861
+ },
+ {
+ "epoch": 21.72,
+ "learning_rate": 0.0001546474358974359,
+ "loss": 0.1347,
+ "step": 862
+ },
+ {
+ "epoch": 21.74,
+ "learning_rate": 0.00015424679487179488,
+ "loss": 0.1322,
+ "step": 863
+ },
+ {
+ "epoch": 21.77,
+ "learning_rate": 0.00015384615384615385,
+ "loss": 0.1321,
+ "step": 864
+ },
+ {
+ "epoch": 21.8,
+ "learning_rate": 0.00015344551282051283,
+ "loss": 0.128,
+ "step": 865
+ },
+ {
+ "epoch": 21.82,
+ "learning_rate": 0.0001530448717948718,
+ "loss": 0.1365,
+ "step": 866
+ },
+ {
+ "epoch": 21.85,
+ "learning_rate": 0.00015264423076923078,
+ "loss": 0.137,
+ "step": 867
+ },
+ {
+ "epoch": 21.87,
+ "learning_rate": 0.00015224358974358972,
+ "loss": 0.1322,
+ "step": 868
+ },
+ {
+ "epoch": 21.9,
+ "learning_rate": 0.0001518429487179487,
+ "loss": 0.1341,
+ "step": 869
+ },
+ {
+ "epoch": 21.92,
+ "learning_rate": 0.00015144230769230767,
+ "loss": 0.1334,
+ "step": 870
+ },
+ {
+ "epoch": 21.95,
+ "learning_rate": 0.00015104166666666665,
+ "loss": 0.1339,
+ "step": 871
+ },
+ {
+ "epoch": 21.97,
+ "learning_rate": 0.00015064102564102565,
+ "loss": 0.1301,
+ "step": 872
+ },
+ {
+ "epoch": 22.0,
+ "learning_rate": 0.00015024038461538462,
+ "loss": 0.1325,
+ "step": 873
+ },
+ {
+ "epoch": 22.02,
+ "learning_rate": 0.0001498397435897436,
+ "loss": 0.1153,
+ "step": 874
+ },
+ {
+ "epoch": 22.05,
+ "learning_rate": 0.00014943910256410257,
+ "loss": 0.1113,
+ "step": 875
+ },
+ {
+ "epoch": 22.07,
+ "learning_rate": 0.00014903846153846155,
+ "loss": 0.1138,
+ "step": 876
+ },
+ {
+ "epoch": 22.1,
+ "learning_rate": 0.00014863782051282052,
+ "loss": 0.1136,
+ "step": 877
+ },
+ {
+ "epoch": 22.12,
+ "learning_rate": 0.0001482371794871795,
+ "loss": 0.1096,
+ "step": 878
+ },
+ {
+ "epoch": 22.15,
+ "learning_rate": 0.00014783653846153847,
+ "loss": 0.116,
+ "step": 879
+ },
+ {
+ "epoch": 22.17,
+ "learning_rate": 0.00014743589743589745,
+ "loss": 0.1122,
+ "step": 880
+ },
+ {
+ "epoch": 22.2,
+ "learning_rate": 0.00014703525641025642,
+ "loss": 0.1132,
+ "step": 881
+ },
+ {
+ "epoch": 22.22,
+ "learning_rate": 0.0001466346153846154,
+ "loss": 0.1128,
+ "step": 882
+ },
+ {
+ "epoch": 22.25,
+ "learning_rate": 0.00014623397435897437,
+ "loss": 0.1134,
+ "step": 883
+ },
+ {
+ "epoch": 22.27,
+ "learning_rate": 0.00014583333333333335,
+ "loss": 0.1065,
+ "step": 884
+ },
+ {
+ "epoch": 22.3,
+ "learning_rate": 0.00014543269230769232,
+ "loss": 0.117,
+ "step": 885
+ },
+ {
+ "epoch": 22.32,
+ "learning_rate": 0.0001450320512820513,
+ "loss": 0.112,
+ "step": 886
+ },
+ {
+ "epoch": 22.35,
+ "learning_rate": 0.00014463141025641027,
+ "loss": 0.1145,
+ "step": 887
+ },
+ {
+ "epoch": 22.37,
+ "learning_rate": 0.00014423076923076922,
+ "loss": 0.1128,
+ "step": 888
+ },
+ {
+ "epoch": 22.4,
+ "learning_rate": 0.0001438301282051282,
+ "loss": 0.1132,
+ "step": 889
+ },
+ {
+ "epoch": 22.43,
+ "learning_rate": 0.00014342948717948717,
+ "loss": 0.1143,
+ "step": 890
+ },
+ {
+ "epoch": 22.45,
+ "learning_rate": 0.00014302884615384614,
+ "loss": 0.1172,
+ "step": 891
+ },
+ {
+ "epoch": 22.48,
+ "learning_rate": 0.00014262820512820512,
+ "loss": 0.1143,
+ "step": 892
+ },
+ {
+ "epoch": 22.5,
+ "learning_rate": 0.0001422275641025641,
+ "loss": 0.1164,
+ "step": 893
+ },
+ {
+ "epoch": 22.53,
+ "learning_rate": 0.00014182692307692307,
+ "loss": 0.1197,
+ "step": 894
+ },
+ {
+ "epoch": 22.55,
+ "learning_rate": 0.00014142628205128204,
+ "loss": 0.1152,
+ "step": 895
+ },
+ {
+ "epoch": 22.58,
+ "learning_rate": 0.00014102564102564101,
+ "loss": 0.1243,
+ "step": 896
+ },
+ {
+ "epoch": 22.6,
+ "learning_rate": 0.00014062500000000002,
+ "loss": 0.1167,
+ "step": 897
+ },
+ {
+ "epoch": 22.63,
+ "learning_rate": 0.000140224358974359,
+ "loss": 0.1224,
+ "step": 898
+ },
+ {
+ "epoch": 22.65,
+ "learning_rate": 0.00013982371794871797,
+ "loss": 0.1192,
+ "step": 899
+ },
+ {
+ "epoch": 22.68,
+ "learning_rate": 0.00013942307692307694,
+ "loss": 0.1184,
+ "step": 900
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 1248,
+ "num_train_epochs": 32,
+ "save_steps": 100,
+ "total_flos": 5.068100806770893e+17,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-900/training_args.bin b/checkpoint-900/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..4fa11fe94c27b6f0864269cde7c6c270e4a135b7
--- /dev/null
+++ b/checkpoint-900/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
+size 4155
diff --git a/training_args.bin b/training_args.bin
index ef063fa00558052539a66e29ecc1b93f5d841a3b..4fa11fe94c27b6f0864269cde7c6c270e4a135b7 100644
--- a/training_args.bin
+++ b/training_args.bin
@@ -1,3 +1,3 @@
version https://git-lfs.github.com/spec/v1
-oid sha256:7b8d3d0c26b3d7e124a186eae5c4ec193759b71982a0e47131ab43fa25d3a439
+oid sha256:7015d16f0f7058d5e16036c5259111dacee395cddfb5fc5ae08d917da3d67f65
size 4155