alykassem commited on
Commit
6225477
·
verified ·
1 Parent(s): 27ba4b2

Training in progress, step 338, checkpoint

Browse files
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ checkpoint-338/tokenizer.json filter=lfs diff=lfs merge=lfs -text
checkpoint-338/chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '
2
+ ' + message['content'] | trim + '<end_of_turn>
3
+ ' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model
4
+ '}}{% endif %}
checkpoint-338/config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "GemmaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 2,
8
+ "eos_token_id": 1,
9
+ "head_dim": 256,
10
+ "hidden_act": "gelu",
11
+ "hidden_activation": null,
12
+ "hidden_size": 2048,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 16384,
15
+ "max_position_embeddings": 2048,
16
+ "model_type": "gemma",
17
+ "num_attention_heads": 8,
18
+ "num_hidden_layers": 18,
19
+ "num_key_value_heads": 1,
20
+ "pad_token_id": 0,
21
+ "rms_norm_eps": 1e-06,
22
+ "rope_scaling": null,
23
+ "rope_theta": 10000.0,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.53.2",
26
+ "unsloth_version": "2024.9",
27
+ "use_cache": true,
28
+ "vocab_size": 256000
29
+ }
checkpoint-338/generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 2,
4
+ "eos_token_id": 1,
5
+ "max_length": 8192,
6
+ "pad_token_id": 0,
7
+ "transformers_version": "4.53.2"
8
+ }
checkpoint-338/model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1ba566c65354efcf1eab09af59fbb80d4a05968117a9be332fcf407ca4da59e
3
+ size 4945242264
checkpoint-338/model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f5b5aed24fc9109803d6c404bf83f0cdb9148412ad63d1a58dea473dce2c7c3
3
+ size 67121608
checkpoint-338/model.safetensors.index.json ADDED
@@ -0,0 +1,172 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_parameters": 2506172416,
4
+ "total_size": 5012344832
5
+ },
6
+ "weight_map": {
7
+ "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00002.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
98
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
99
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
100
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
101
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
102
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
103
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
104
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
105
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
106
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
107
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
108
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
109
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
110
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
111
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
112
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
113
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
114
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
115
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
116
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
117
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
118
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
119
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
120
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
121
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
122
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
123
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
124
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
125
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
126
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
127
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
128
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
129
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
130
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
131
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
132
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
133
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
134
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
135
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
136
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
137
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
138
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
139
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
140
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
141
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
142
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
143
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors",
144
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
145
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
146
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
147
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
148
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
149
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
150
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
151
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
152
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors",
153
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
154
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
155
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
156
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
157
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
158
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
159
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
160
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
161
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors",
162
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
163
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
164
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
165
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
166
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
167
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
168
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
169
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
170
+ "model.norm.weight": "model-00002-of-00002.safetensors"
171
+ }
172
+ }
checkpoint-338/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:625b89b528343136b1c7c9a27154b59e2e5da03f55aa4ce4551c6ed832a636bb
3
+ size 8220638802
checkpoint-338/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e271db059ee705520b0102bdbb522c5c372acf668aa9b12f38477021ca2f6b79
3
+ size 14180
checkpoint-338/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f1645fc6c6769deebd7949e8908f6796c92fdfac4038f3346c9b63182736efd
3
+ size 1064
checkpoint-338/special_tokens_map.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<start_of_turn>",
4
+ "<end_of_turn>"
5
+ ],
6
+ "bos_token": {
7
+ "content": "<bos>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "eos_token": {
14
+ "content": "<eos>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "pad_token": "<eos>",
21
+ "unk_token": {
22
+ "content": "<unk>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false
27
+ }
28
+ }
checkpoint-338/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f559f2189f392b4555613965f089e7c4d300b41fbe080bf79da0d676e33ee7f0
3
+ size 34356041
checkpoint-338/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61a7b147390c64585d6c3543dd6fc636906c9af3865a5548f27f31aee1d4c8e2
3
+ size 4241003
checkpoint-338/tokenizer_config.json ADDED
@@ -0,0 +1,1758 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<eos>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "<bos>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "<unk>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "4": {
38
+ "content": "<mask>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": false
44
+ },
45
+ "5": {
46
+ "content": "<2mass>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": false
52
+ },
53
+ "6": {
54
+ "content": "[@BOS@]",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": false
60
+ },
61
+ "7": {
62
+ "content": "<unused0>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": false
68
+ },
69
+ "8": {
70
+ "content": "<unused1>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": false
76
+ },
77
+ "9": {
78
+ "content": "<unused2>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": false
84
+ },
85
+ "10": {
86
+ "content": "<unused3>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": false
92
+ },
93
+ "11": {
94
+ "content": "<unused4>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": false
100
+ },
101
+ "12": {
102
+ "content": "<unused5>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": false
108
+ },
109
+ "13": {
110
+ "content": "<unused6>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": false
116
+ },
117
+ "14": {
118
+ "content": "<unused7>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "15": {
126
+ "content": "<unused8>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "16": {
134
+ "content": "<unused9>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "17": {
142
+ "content": "<unused10>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "18": {
150
+ "content": "<unused11>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "19": {
158
+ "content": "<unused12>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "20": {
166
+ "content": "<unused13>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "21": {
174
+ "content": "<unused14>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "22": {
182
+ "content": "<unused15>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "23": {
190
+ "content": "<unused16>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "24": {
198
+ "content": "<unused17>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "25": {
206
+ "content": "<unused18>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ },
213
+ "26": {
214
+ "content": "<unused19>",
215
+ "lstrip": false,
216
+ "normalized": false,
217
+ "rstrip": false,
218
+ "single_word": false,
219
+ "special": false
220
+ },
221
+ "27": {
222
+ "content": "<unused20>",
223
+ "lstrip": false,
224
+ "normalized": false,
225
+ "rstrip": false,
226
+ "single_word": false,
227
+ "special": false
228
+ },
229
+ "28": {
230
+ "content": "<unused21>",
231
+ "lstrip": false,
232
+ "normalized": false,
233
+ "rstrip": false,
234
+ "single_word": false,
235
+ "special": false
236
+ },
237
+ "29": {
238
+ "content": "<unused22>",
239
+ "lstrip": false,
240
+ "normalized": false,
241
+ "rstrip": false,
242
+ "single_word": false,
243
+ "special": false
244
+ },
245
+ "30": {
246
+ "content": "<unused23>",
247
+ "lstrip": false,
248
+ "normalized": false,
249
+ "rstrip": false,
250
+ "single_word": false,
251
+ "special": false
252
+ },
253
+ "31": {
254
+ "content": "<unused24>",
255
+ "lstrip": false,
256
+ "normalized": false,
257
+ "rstrip": false,
258
+ "single_word": false,
259
+ "special": false
260
+ },
261
+ "32": {
262
+ "content": "<unused25>",
263
+ "lstrip": false,
264
+ "normalized": false,
265
+ "rstrip": false,
266
+ "single_word": false,
267
+ "special": false
268
+ },
269
+ "33": {
270
+ "content": "<unused26>",
271
+ "lstrip": false,
272
+ "normalized": false,
273
+ "rstrip": false,
274
+ "single_word": false,
275
+ "special": false
276
+ },
277
+ "34": {
278
+ "content": "<unused27>",
279
+ "lstrip": false,
280
+ "normalized": false,
281
+ "rstrip": false,
282
+ "single_word": false,
283
+ "special": false
284
+ },
285
+ "35": {
286
+ "content": "<unused28>",
287
+ "lstrip": false,
288
+ "normalized": false,
289
+ "rstrip": false,
290
+ "single_word": false,
291
+ "special": false
292
+ },
293
+ "36": {
294
+ "content": "<unused29>",
295
+ "lstrip": false,
296
+ "normalized": false,
297
+ "rstrip": false,
298
+ "single_word": false,
299
+ "special": false
300
+ },
301
+ "37": {
302
+ "content": "<unused30>",
303
+ "lstrip": false,
304
+ "normalized": false,
305
+ "rstrip": false,
306
+ "single_word": false,
307
+ "special": false
308
+ },
309
+ "38": {
310
+ "content": "<unused31>",
311
+ "lstrip": false,
312
+ "normalized": false,
313
+ "rstrip": false,
314
+ "single_word": false,
315
+ "special": false
316
+ },
317
+ "39": {
318
+ "content": "<unused32>",
319
+ "lstrip": false,
320
+ "normalized": false,
321
+ "rstrip": false,
322
+ "single_word": false,
323
+ "special": false
324
+ },
325
+ "40": {
326
+ "content": "<unused33>",
327
+ "lstrip": false,
328
+ "normalized": false,
329
+ "rstrip": false,
330
+ "single_word": false,
331
+ "special": false
332
+ },
333
+ "41": {
334
+ "content": "<unused34>",
335
+ "lstrip": false,
336
+ "normalized": false,
337
+ "rstrip": false,
338
+ "single_word": false,
339
+ "special": false
340
+ },
341
+ "42": {
342
+ "content": "<unused35>",
343
+ "lstrip": false,
344
+ "normalized": false,
345
+ "rstrip": false,
346
+ "single_word": false,
347
+ "special": false
348
+ },
349
+ "43": {
350
+ "content": "<unused36>",
351
+ "lstrip": false,
352
+ "normalized": false,
353
+ "rstrip": false,
354
+ "single_word": false,
355
+ "special": false
356
+ },
357
+ "44": {
358
+ "content": "<unused37>",
359
+ "lstrip": false,
360
+ "normalized": false,
361
+ "rstrip": false,
362
+ "single_word": false,
363
+ "special": false
364
+ },
365
+ "45": {
366
+ "content": "<unused38>",
367
+ "lstrip": false,
368
+ "normalized": false,
369
+ "rstrip": false,
370
+ "single_word": false,
371
+ "special": false
372
+ },
373
+ "46": {
374
+ "content": "<unused39>",
375
+ "lstrip": false,
376
+ "normalized": false,
377
+ "rstrip": false,
378
+ "single_word": false,
379
+ "special": false
380
+ },
381
+ "47": {
382
+ "content": "<unused40>",
383
+ "lstrip": false,
384
+ "normalized": false,
385
+ "rstrip": false,
386
+ "single_word": false,
387
+ "special": false
388
+ },
389
+ "48": {
390
+ "content": "<unused41>",
391
+ "lstrip": false,
392
+ "normalized": false,
393
+ "rstrip": false,
394
+ "single_word": false,
395
+ "special": false
396
+ },
397
+ "49": {
398
+ "content": "<unused42>",
399
+ "lstrip": false,
400
+ "normalized": false,
401
+ "rstrip": false,
402
+ "single_word": false,
403
+ "special": false
404
+ },
405
+ "50": {
406
+ "content": "<unused43>",
407
+ "lstrip": false,
408
+ "normalized": false,
409
+ "rstrip": false,
410
+ "single_word": false,
411
+ "special": false
412
+ },
413
+ "51": {
414
+ "content": "<unused44>",
415
+ "lstrip": false,
416
+ "normalized": false,
417
+ "rstrip": false,
418
+ "single_word": false,
419
+ "special": false
420
+ },
421
+ "52": {
422
+ "content": "<unused45>",
423
+ "lstrip": false,
424
+ "normalized": false,
425
+ "rstrip": false,
426
+ "single_word": false,
427
+ "special": false
428
+ },
429
+ "53": {
430
+ "content": "<unused46>",
431
+ "lstrip": false,
432
+ "normalized": false,
433
+ "rstrip": false,
434
+ "single_word": false,
435
+ "special": false
436
+ },
437
+ "54": {
438
+ "content": "<unused47>",
439
+ "lstrip": false,
440
+ "normalized": false,
441
+ "rstrip": false,
442
+ "single_word": false,
443
+ "special": false
444
+ },
445
+ "55": {
446
+ "content": "<unused48>",
447
+ "lstrip": false,
448
+ "normalized": false,
449
+ "rstrip": false,
450
+ "single_word": false,
451
+ "special": false
452
+ },
453
+ "56": {
454
+ "content": "<unused49>",
455
+ "lstrip": false,
456
+ "normalized": false,
457
+ "rstrip": false,
458
+ "single_word": false,
459
+ "special": false
460
+ },
461
+ "57": {
462
+ "content": "<unused50>",
463
+ "lstrip": false,
464
+ "normalized": false,
465
+ "rstrip": false,
466
+ "single_word": false,
467
+ "special": false
468
+ },
469
+ "58": {
470
+ "content": "<unused51>",
471
+ "lstrip": false,
472
+ "normalized": false,
473
+ "rstrip": false,
474
+ "single_word": false,
475
+ "special": false
476
+ },
477
+ "59": {
478
+ "content": "<unused52>",
479
+ "lstrip": false,
480
+ "normalized": false,
481
+ "rstrip": false,
482
+ "single_word": false,
483
+ "special": false
484
+ },
485
+ "60": {
486
+ "content": "<unused53>",
487
+ "lstrip": false,
488
+ "normalized": false,
489
+ "rstrip": false,
490
+ "single_word": false,
491
+ "special": false
492
+ },
493
+ "61": {
494
+ "content": "<unused54>",
495
+ "lstrip": false,
496
+ "normalized": false,
497
+ "rstrip": false,
498
+ "single_word": false,
499
+ "special": false
500
+ },
501
+ "62": {
502
+ "content": "<unused55>",
503
+ "lstrip": false,
504
+ "normalized": false,
505
+ "rstrip": false,
506
+ "single_word": false,
507
+ "special": false
508
+ },
509
+ "63": {
510
+ "content": "<unused56>",
511
+ "lstrip": false,
512
+ "normalized": false,
513
+ "rstrip": false,
514
+ "single_word": false,
515
+ "special": false
516
+ },
517
+ "64": {
518
+ "content": "<unused57>",
519
+ "lstrip": false,
520
+ "normalized": false,
521
+ "rstrip": false,
522
+ "single_word": false,
523
+ "special": false
524
+ },
525
+ "65": {
526
+ "content": "<unused58>",
527
+ "lstrip": false,
528
+ "normalized": false,
529
+ "rstrip": false,
530
+ "single_word": false,
531
+ "special": false
532
+ },
533
+ "66": {
534
+ "content": "<unused59>",
535
+ "lstrip": false,
536
+ "normalized": false,
537
+ "rstrip": false,
538
+ "single_word": false,
539
+ "special": false
540
+ },
541
+ "67": {
542
+ "content": "<unused60>",
543
+ "lstrip": false,
544
+ "normalized": false,
545
+ "rstrip": false,
546
+ "single_word": false,
547
+ "special": false
548
+ },
549
+ "68": {
550
+ "content": "<unused61>",
551
+ "lstrip": false,
552
+ "normalized": false,
553
+ "rstrip": false,
554
+ "single_word": false,
555
+ "special": false
556
+ },
557
+ "69": {
558
+ "content": "<unused62>",
559
+ "lstrip": false,
560
+ "normalized": false,
561
+ "rstrip": false,
562
+ "single_word": false,
563
+ "special": false
564
+ },
565
+ "70": {
566
+ "content": "<unused63>",
567
+ "lstrip": false,
568
+ "normalized": false,
569
+ "rstrip": false,
570
+ "single_word": false,
571
+ "special": false
572
+ },
573
+ "71": {
574
+ "content": "<unused64>",
575
+ "lstrip": false,
576
+ "normalized": false,
577
+ "rstrip": false,
578
+ "single_word": false,
579
+ "special": false
580
+ },
581
+ "72": {
582
+ "content": "<unused65>",
583
+ "lstrip": false,
584
+ "normalized": false,
585
+ "rstrip": false,
586
+ "single_word": false,
587
+ "special": false
588
+ },
589
+ "73": {
590
+ "content": "<unused66>",
591
+ "lstrip": false,
592
+ "normalized": false,
593
+ "rstrip": false,
594
+ "single_word": false,
595
+ "special": false
596
+ },
597
+ "74": {
598
+ "content": "<unused67>",
599
+ "lstrip": false,
600
+ "normalized": false,
601
+ "rstrip": false,
602
+ "single_word": false,
603
+ "special": false
604
+ },
605
+ "75": {
606
+ "content": "<unused68>",
607
+ "lstrip": false,
608
+ "normalized": false,
609
+ "rstrip": false,
610
+ "single_word": false,
611
+ "special": false
612
+ },
613
+ "76": {
614
+ "content": "<unused69>",
615
+ "lstrip": false,
616
+ "normalized": false,
617
+ "rstrip": false,
618
+ "single_word": false,
619
+ "special": false
620
+ },
621
+ "77": {
622
+ "content": "<unused70>",
623
+ "lstrip": false,
624
+ "normalized": false,
625
+ "rstrip": false,
626
+ "single_word": false,
627
+ "special": false
628
+ },
629
+ "78": {
630
+ "content": "<unused71>",
631
+ "lstrip": false,
632
+ "normalized": false,
633
+ "rstrip": false,
634
+ "single_word": false,
635
+ "special": false
636
+ },
637
+ "79": {
638
+ "content": "<unused72>",
639
+ "lstrip": false,
640
+ "normalized": false,
641
+ "rstrip": false,
642
+ "single_word": false,
643
+ "special": false
644
+ },
645
+ "80": {
646
+ "content": "<unused73>",
647
+ "lstrip": false,
648
+ "normalized": false,
649
+ "rstrip": false,
650
+ "single_word": false,
651
+ "special": false
652
+ },
653
+ "81": {
654
+ "content": "<unused74>",
655
+ "lstrip": false,
656
+ "normalized": false,
657
+ "rstrip": false,
658
+ "single_word": false,
659
+ "special": false
660
+ },
661
+ "82": {
662
+ "content": "<unused75>",
663
+ "lstrip": false,
664
+ "normalized": false,
665
+ "rstrip": false,
666
+ "single_word": false,
667
+ "special": false
668
+ },
669
+ "83": {
670
+ "content": "<unused76>",
671
+ "lstrip": false,
672
+ "normalized": false,
673
+ "rstrip": false,
674
+ "single_word": false,
675
+ "special": false
676
+ },
677
+ "84": {
678
+ "content": "<unused77>",
679
+ "lstrip": false,
680
+ "normalized": false,
681
+ "rstrip": false,
682
+ "single_word": false,
683
+ "special": false
684
+ },
685
+ "85": {
686
+ "content": "<unused78>",
687
+ "lstrip": false,
688
+ "normalized": false,
689
+ "rstrip": false,
690
+ "single_word": false,
691
+ "special": false
692
+ },
693
+ "86": {
694
+ "content": "<unused79>",
695
+ "lstrip": false,
696
+ "normalized": false,
697
+ "rstrip": false,
698
+ "single_word": false,
699
+ "special": false
700
+ },
701
+ "87": {
702
+ "content": "<unused80>",
703
+ "lstrip": false,
704
+ "normalized": false,
705
+ "rstrip": false,
706
+ "single_word": false,
707
+ "special": false
708
+ },
709
+ "88": {
710
+ "content": "<unused81>",
711
+ "lstrip": false,
712
+ "normalized": false,
713
+ "rstrip": false,
714
+ "single_word": false,
715
+ "special": false
716
+ },
717
+ "89": {
718
+ "content": "<unused82>",
719
+ "lstrip": false,
720
+ "normalized": false,
721
+ "rstrip": false,
722
+ "single_word": false,
723
+ "special": false
724
+ },
725
+ "90": {
726
+ "content": "<unused83>",
727
+ "lstrip": false,
728
+ "normalized": false,
729
+ "rstrip": false,
730
+ "single_word": false,
731
+ "special": false
732
+ },
733
+ "91": {
734
+ "content": "<unused84>",
735
+ "lstrip": false,
736
+ "normalized": false,
737
+ "rstrip": false,
738
+ "single_word": false,
739
+ "special": false
740
+ },
741
+ "92": {
742
+ "content": "<unused85>",
743
+ "lstrip": false,
744
+ "normalized": false,
745
+ "rstrip": false,
746
+ "single_word": false,
747
+ "special": false
748
+ },
749
+ "93": {
750
+ "content": "<unused86>",
751
+ "lstrip": false,
752
+ "normalized": false,
753
+ "rstrip": false,
754
+ "single_word": false,
755
+ "special": false
756
+ },
757
+ "94": {
758
+ "content": "<unused87>",
759
+ "lstrip": false,
760
+ "normalized": false,
761
+ "rstrip": false,
762
+ "single_word": false,
763
+ "special": false
764
+ },
765
+ "95": {
766
+ "content": "<unused88>",
767
+ "lstrip": false,
768
+ "normalized": false,
769
+ "rstrip": false,
770
+ "single_word": false,
771
+ "special": false
772
+ },
773
+ "96": {
774
+ "content": "<unused89>",
775
+ "lstrip": false,
776
+ "normalized": false,
777
+ "rstrip": false,
778
+ "single_word": false,
779
+ "special": false
780
+ },
781
+ "97": {
782
+ "content": "<unused90>",
783
+ "lstrip": false,
784
+ "normalized": false,
785
+ "rstrip": false,
786
+ "single_word": false,
787
+ "special": false
788
+ },
789
+ "98": {
790
+ "content": "<unused91>",
791
+ "lstrip": false,
792
+ "normalized": false,
793
+ "rstrip": false,
794
+ "single_word": false,
795
+ "special": false
796
+ },
797
+ "99": {
798
+ "content": "<unused92>",
799
+ "lstrip": false,
800
+ "normalized": false,
801
+ "rstrip": false,
802
+ "single_word": false,
803
+ "special": false
804
+ },
805
+ "100": {
806
+ "content": "<unused93>",
807
+ "lstrip": false,
808
+ "normalized": false,
809
+ "rstrip": false,
810
+ "single_word": false,
811
+ "special": false
812
+ },
813
+ "101": {
814
+ "content": "<unused94>",
815
+ "lstrip": false,
816
+ "normalized": false,
817
+ "rstrip": false,
818
+ "single_word": false,
819
+ "special": false
820
+ },
821
+ "102": {
822
+ "content": "<unused95>",
823
+ "lstrip": false,
824
+ "normalized": false,
825
+ "rstrip": false,
826
+ "single_word": false,
827
+ "special": false
828
+ },
829
+ "103": {
830
+ "content": "<unused96>",
831
+ "lstrip": false,
832
+ "normalized": false,
833
+ "rstrip": false,
834
+ "single_word": false,
835
+ "special": false
836
+ },
837
+ "104": {
838
+ "content": "<unused97>",
839
+ "lstrip": false,
840
+ "normalized": false,
841
+ "rstrip": false,
842
+ "single_word": false,
843
+ "special": false
844
+ },
845
+ "105": {
846
+ "content": "<unused98>",
847
+ "lstrip": false,
848
+ "normalized": false,
849
+ "rstrip": false,
850
+ "single_word": false,
851
+ "special": false
852
+ },
853
+ "106": {
854
+ "content": "<start_of_turn>",
855
+ "lstrip": false,
856
+ "normalized": false,
857
+ "rstrip": false,
858
+ "single_word": false,
859
+ "special": true
860
+ },
861
+ "107": {
862
+ "content": "<end_of_turn>",
863
+ "lstrip": false,
864
+ "normalized": false,
865
+ "rstrip": false,
866
+ "single_word": false,
867
+ "special": true
868
+ },
869
+ "108": {
870
+ "content": "\n",
871
+ "lstrip": false,
872
+ "normalized": false,
873
+ "rstrip": false,
874
+ "single_word": false,
875
+ "special": false
876
+ },
877
+ "109": {
878
+ "content": "\n\n",
879
+ "lstrip": false,
880
+ "normalized": false,
881
+ "rstrip": false,
882
+ "single_word": false,
883
+ "special": false
884
+ },
885
+ "110": {
886
+ "content": "\n\n\n",
887
+ "lstrip": false,
888
+ "normalized": false,
889
+ "rstrip": false,
890
+ "single_word": false,
891
+ "special": false
892
+ },
893
+ "111": {
894
+ "content": "\n\n\n\n",
895
+ "lstrip": false,
896
+ "normalized": false,
897
+ "rstrip": false,
898
+ "single_word": false,
899
+ "special": false
900
+ },
901
+ "112": {
902
+ "content": "\n\n\n\n\n",
903
+ "lstrip": false,
904
+ "normalized": false,
905
+ "rstrip": false,
906
+ "single_word": false,
907
+ "special": false
908
+ },
909
+ "113": {
910
+ "content": "\n\n\n\n\n\n",
911
+ "lstrip": false,
912
+ "normalized": false,
913
+ "rstrip": false,
914
+ "single_word": false,
915
+ "special": false
916
+ },
917
+ "114": {
918
+ "content": "\n\n\n\n\n\n\n",
919
+ "lstrip": false,
920
+ "normalized": false,
921
+ "rstrip": false,
922
+ "single_word": false,
923
+ "special": false
924
+ },
925
+ "115": {
926
+ "content": "\n\n\n\n\n\n\n\n",
927
+ "lstrip": false,
928
+ "normalized": false,
929
+ "rstrip": false,
930
+ "single_word": false,
931
+ "special": false
932
+ },
933
+ "116": {
934
+ "content": "\n\n\n\n\n\n\n\n\n",
935
+ "lstrip": false,
936
+ "normalized": false,
937
+ "rstrip": false,
938
+ "single_word": false,
939
+ "special": false
940
+ },
941
+ "117": {
942
+ "content": "\n\n\n\n\n\n\n\n\n\n",
943
+ "lstrip": false,
944
+ "normalized": false,
945
+ "rstrip": false,
946
+ "single_word": false,
947
+ "special": false
948
+ },
949
+ "118": {
950
+ "content": "\n\n\n\n\n\n\n\n\n\n\n",
951
+ "lstrip": false,
952
+ "normalized": false,
953
+ "rstrip": false,
954
+ "single_word": false,
955
+ "special": false
956
+ },
957
+ "119": {
958
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n",
959
+ "lstrip": false,
960
+ "normalized": false,
961
+ "rstrip": false,
962
+ "single_word": false,
963
+ "special": false
964
+ },
965
+ "120": {
966
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n",
967
+ "lstrip": false,
968
+ "normalized": false,
969
+ "rstrip": false,
970
+ "single_word": false,
971
+ "special": false
972
+ },
973
+ "121": {
974
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
975
+ "lstrip": false,
976
+ "normalized": false,
977
+ "rstrip": false,
978
+ "single_word": false,
979
+ "special": false
980
+ },
981
+ "122": {
982
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
983
+ "lstrip": false,
984
+ "normalized": false,
985
+ "rstrip": false,
986
+ "single_word": false,
987
+ "special": false
988
+ },
989
+ "123": {
990
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
991
+ "lstrip": false,
992
+ "normalized": false,
993
+ "rstrip": false,
994
+ "single_word": false,
995
+ "special": false
996
+ },
997
+ "124": {
998
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
999
+ "lstrip": false,
1000
+ "normalized": false,
1001
+ "rstrip": false,
1002
+ "single_word": false,
1003
+ "special": false
1004
+ },
1005
+ "125": {
1006
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1007
+ "lstrip": false,
1008
+ "normalized": false,
1009
+ "rstrip": false,
1010
+ "single_word": false,
1011
+ "special": false
1012
+ },
1013
+ "126": {
1014
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1015
+ "lstrip": false,
1016
+ "normalized": false,
1017
+ "rstrip": false,
1018
+ "single_word": false,
1019
+ "special": false
1020
+ },
1021
+ "127": {
1022
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1023
+ "lstrip": false,
1024
+ "normalized": false,
1025
+ "rstrip": false,
1026
+ "single_word": false,
1027
+ "special": false
1028
+ },
1029
+ "128": {
1030
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1031
+ "lstrip": false,
1032
+ "normalized": false,
1033
+ "rstrip": false,
1034
+ "single_word": false,
1035
+ "special": false
1036
+ },
1037
+ "129": {
1038
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1039
+ "lstrip": false,
1040
+ "normalized": false,
1041
+ "rstrip": false,
1042
+ "single_word": false,
1043
+ "special": false
1044
+ },
1045
+ "130": {
1046
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1047
+ "lstrip": false,
1048
+ "normalized": false,
1049
+ "rstrip": false,
1050
+ "single_word": false,
1051
+ "special": false
1052
+ },
1053
+ "131": {
1054
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1055
+ "lstrip": false,
1056
+ "normalized": false,
1057
+ "rstrip": false,
1058
+ "single_word": false,
1059
+ "special": false
1060
+ },
1061
+ "132": {
1062
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1063
+ "lstrip": false,
1064
+ "normalized": false,
1065
+ "rstrip": false,
1066
+ "single_word": false,
1067
+ "special": false
1068
+ },
1069
+ "133": {
1070
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1071
+ "lstrip": false,
1072
+ "normalized": false,
1073
+ "rstrip": false,
1074
+ "single_word": false,
1075
+ "special": false
1076
+ },
1077
+ "134": {
1078
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1079
+ "lstrip": false,
1080
+ "normalized": false,
1081
+ "rstrip": false,
1082
+ "single_word": false,
1083
+ "special": false
1084
+ },
1085
+ "135": {
1086
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1087
+ "lstrip": false,
1088
+ "normalized": false,
1089
+ "rstrip": false,
1090
+ "single_word": false,
1091
+ "special": false
1092
+ },
1093
+ "136": {
1094
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1095
+ "lstrip": false,
1096
+ "normalized": false,
1097
+ "rstrip": false,
1098
+ "single_word": false,
1099
+ "special": false
1100
+ },
1101
+ "137": {
1102
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1103
+ "lstrip": false,
1104
+ "normalized": false,
1105
+ "rstrip": false,
1106
+ "single_word": false,
1107
+ "special": false
1108
+ },
1109
+ "138": {
1110
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1111
+ "lstrip": false,
1112
+ "normalized": false,
1113
+ "rstrip": false,
1114
+ "single_word": false,
1115
+ "special": false
1116
+ },
1117
+ "139": {
1118
+ "content": "▁▁",
1119
+ "lstrip": false,
1120
+ "normalized": false,
1121
+ "rstrip": false,
1122
+ "single_word": false,
1123
+ "special": false
1124
+ },
1125
+ "140": {
1126
+ "content": "▁▁▁",
1127
+ "lstrip": false,
1128
+ "normalized": false,
1129
+ "rstrip": false,
1130
+ "single_word": false,
1131
+ "special": false
1132
+ },
1133
+ "141": {
1134
+ "content": "▁▁▁▁",
1135
+ "lstrip": false,
1136
+ "normalized": false,
1137
+ "rstrip": false,
1138
+ "single_word": false,
1139
+ "special": false
1140
+ },
1141
+ "142": {
1142
+ "content": "▁▁▁▁▁",
1143
+ "lstrip": false,
1144
+ "normalized": false,
1145
+ "rstrip": false,
1146
+ "single_word": false,
1147
+ "special": false
1148
+ },
1149
+ "143": {
1150
+ "content": "▁▁▁▁▁▁",
1151
+ "lstrip": false,
1152
+ "normalized": false,
1153
+ "rstrip": false,
1154
+ "single_word": false,
1155
+ "special": false
1156
+ },
1157
+ "144": {
1158
+ "content": "▁▁▁▁▁▁▁",
1159
+ "lstrip": false,
1160
+ "normalized": false,
1161
+ "rstrip": false,
1162
+ "single_word": false,
1163
+ "special": false
1164
+ },
1165
+ "145": {
1166
+ "content": "▁▁▁▁▁▁▁▁",
1167
+ "lstrip": false,
1168
+ "normalized": false,
1169
+ "rstrip": false,
1170
+ "single_word": false,
1171
+ "special": false
1172
+ },
1173
+ "146": {
1174
+ "content": "▁▁▁▁▁▁▁▁▁",
1175
+ "lstrip": false,
1176
+ "normalized": false,
1177
+ "rstrip": false,
1178
+ "single_word": false,
1179
+ "special": false
1180
+ },
1181
+ "147": {
1182
+ "content": "▁▁▁▁▁▁▁▁▁▁",
1183
+ "lstrip": false,
1184
+ "normalized": false,
1185
+ "rstrip": false,
1186
+ "single_word": false,
1187
+ "special": false
1188
+ },
1189
+ "148": {
1190
+ "content": "▁▁▁▁▁▁▁▁▁▁▁",
1191
+ "lstrip": false,
1192
+ "normalized": false,
1193
+ "rstrip": false,
1194
+ "single_word": false,
1195
+ "special": false
1196
+ },
1197
+ "149": {
1198
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁",
1199
+ "lstrip": false,
1200
+ "normalized": false,
1201
+ "rstrip": false,
1202
+ "single_word": false,
1203
+ "special": false
1204
+ },
1205
+ "150": {
1206
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁",
1207
+ "lstrip": false,
1208
+ "normalized": false,
1209
+ "rstrip": false,
1210
+ "single_word": false,
1211
+ "special": false
1212
+ },
1213
+ "151": {
1214
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1215
+ "lstrip": false,
1216
+ "normalized": false,
1217
+ "rstrip": false,
1218
+ "single_word": false,
1219
+ "special": false
1220
+ },
1221
+ "152": {
1222
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1223
+ "lstrip": false,
1224
+ "normalized": false,
1225
+ "rstrip": false,
1226
+ "single_word": false,
1227
+ "special": false
1228
+ },
1229
+ "153": {
1230
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1231
+ "lstrip": false,
1232
+ "normalized": false,
1233
+ "rstrip": false,
1234
+ "single_word": false,
1235
+ "special": false
1236
+ },
1237
+ "154": {
1238
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1239
+ "lstrip": false,
1240
+ "normalized": false,
1241
+ "rstrip": false,
1242
+ "single_word": false,
1243
+ "special": false
1244
+ },
1245
+ "155": {
1246
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1247
+ "lstrip": false,
1248
+ "normalized": false,
1249
+ "rstrip": false,
1250
+ "single_word": false,
1251
+ "special": false
1252
+ },
1253
+ "156": {
1254
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1255
+ "lstrip": false,
1256
+ "normalized": false,
1257
+ "rstrip": false,
1258
+ "single_word": false,
1259
+ "special": false
1260
+ },
1261
+ "157": {
1262
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1263
+ "lstrip": false,
1264
+ "normalized": false,
1265
+ "rstrip": false,
1266
+ "single_word": false,
1267
+ "special": false
1268
+ },
1269
+ "158": {
1270
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1271
+ "lstrip": false,
1272
+ "normalized": false,
1273
+ "rstrip": false,
1274
+ "single_word": false,
1275
+ "special": false
1276
+ },
1277
+ "159": {
1278
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1279
+ "lstrip": false,
1280
+ "normalized": false,
1281
+ "rstrip": false,
1282
+ "single_word": false,
1283
+ "special": false
1284
+ },
1285
+ "160": {
1286
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1287
+ "lstrip": false,
1288
+ "normalized": false,
1289
+ "rstrip": false,
1290
+ "single_word": false,
1291
+ "special": false
1292
+ },
1293
+ "161": {
1294
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1295
+ "lstrip": false,
1296
+ "normalized": false,
1297
+ "rstrip": false,
1298
+ "single_word": false,
1299
+ "special": false
1300
+ },
1301
+ "162": {
1302
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1303
+ "lstrip": false,
1304
+ "normalized": false,
1305
+ "rstrip": false,
1306
+ "single_word": false,
1307
+ "special": false
1308
+ },
1309
+ "163": {
1310
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1311
+ "lstrip": false,
1312
+ "normalized": false,
1313
+ "rstrip": false,
1314
+ "single_word": false,
1315
+ "special": false
1316
+ },
1317
+ "164": {
1318
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1319
+ "lstrip": false,
1320
+ "normalized": false,
1321
+ "rstrip": false,
1322
+ "single_word": false,
1323
+ "special": false
1324
+ },
1325
+ "165": {
1326
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1327
+ "lstrip": false,
1328
+ "normalized": false,
1329
+ "rstrip": false,
1330
+ "single_word": false,
1331
+ "special": false
1332
+ },
1333
+ "166": {
1334
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1335
+ "lstrip": false,
1336
+ "normalized": false,
1337
+ "rstrip": false,
1338
+ "single_word": false,
1339
+ "special": false
1340
+ },
1341
+ "167": {
1342
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1343
+ "lstrip": false,
1344
+ "normalized": false,
1345
+ "rstrip": false,
1346
+ "single_word": false,
1347
+ "special": false
1348
+ },
1349
+ "168": {
1350
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1351
+ "lstrip": false,
1352
+ "normalized": false,
1353
+ "rstrip": false,
1354
+ "single_word": false,
1355
+ "special": false
1356
+ },
1357
+ "169": {
1358
+ "content": "<table>",
1359
+ "lstrip": false,
1360
+ "normalized": false,
1361
+ "rstrip": false,
1362
+ "single_word": false,
1363
+ "special": false
1364
+ },
1365
+ "170": {
1366
+ "content": "<caption>",
1367
+ "lstrip": false,
1368
+ "normalized": false,
1369
+ "rstrip": false,
1370
+ "single_word": false,
1371
+ "special": false
1372
+ },
1373
+ "171": {
1374
+ "content": "<thead>",
1375
+ "lstrip": false,
1376
+ "normalized": false,
1377
+ "rstrip": false,
1378
+ "single_word": false,
1379
+ "special": false
1380
+ },
1381
+ "172": {
1382
+ "content": "<tbody>",
1383
+ "lstrip": false,
1384
+ "normalized": false,
1385
+ "rstrip": false,
1386
+ "single_word": false,
1387
+ "special": false
1388
+ },
1389
+ "173": {
1390
+ "content": "<tfoot>",
1391
+ "lstrip": false,
1392
+ "normalized": false,
1393
+ "rstrip": false,
1394
+ "single_word": false,
1395
+ "special": false
1396
+ },
1397
+ "174": {
1398
+ "content": "<tr>",
1399
+ "lstrip": false,
1400
+ "normalized": false,
1401
+ "rstrip": false,
1402
+ "single_word": false,
1403
+ "special": false
1404
+ },
1405
+ "175": {
1406
+ "content": "<th>",
1407
+ "lstrip": false,
1408
+ "normalized": false,
1409
+ "rstrip": false,
1410
+ "single_word": false,
1411
+ "special": false
1412
+ },
1413
+ "176": {
1414
+ "content": "<td>",
1415
+ "lstrip": false,
1416
+ "normalized": false,
1417
+ "rstrip": false,
1418
+ "single_word": false,
1419
+ "special": false
1420
+ },
1421
+ "177": {
1422
+ "content": "</table>",
1423
+ "lstrip": false,
1424
+ "normalized": false,
1425
+ "rstrip": false,
1426
+ "single_word": false,
1427
+ "special": false
1428
+ },
1429
+ "178": {
1430
+ "content": "</caption>",
1431
+ "lstrip": false,
1432
+ "normalized": false,
1433
+ "rstrip": false,
1434
+ "single_word": false,
1435
+ "special": false
1436
+ },
1437
+ "179": {
1438
+ "content": "</thead>",
1439
+ "lstrip": false,
1440
+ "normalized": false,
1441
+ "rstrip": false,
1442
+ "single_word": false,
1443
+ "special": false
1444
+ },
1445
+ "180": {
1446
+ "content": "</tbody>",
1447
+ "lstrip": false,
1448
+ "normalized": false,
1449
+ "rstrip": false,
1450
+ "single_word": false,
1451
+ "special": false
1452
+ },
1453
+ "181": {
1454
+ "content": "</tfoot>",
1455
+ "lstrip": false,
1456
+ "normalized": false,
1457
+ "rstrip": false,
1458
+ "single_word": false,
1459
+ "special": false
1460
+ },
1461
+ "182": {
1462
+ "content": "</tr>",
1463
+ "lstrip": false,
1464
+ "normalized": false,
1465
+ "rstrip": false,
1466
+ "single_word": false,
1467
+ "special": false
1468
+ },
1469
+ "183": {
1470
+ "content": "</th>",
1471
+ "lstrip": false,
1472
+ "normalized": false,
1473
+ "rstrip": false,
1474
+ "single_word": false,
1475
+ "special": false
1476
+ },
1477
+ "184": {
1478
+ "content": "</td>",
1479
+ "lstrip": false,
1480
+ "normalized": false,
1481
+ "rstrip": false,
1482
+ "single_word": false,
1483
+ "special": false
1484
+ },
1485
+ "185": {
1486
+ "content": "<h1>",
1487
+ "lstrip": false,
1488
+ "normalized": false,
1489
+ "rstrip": false,
1490
+ "single_word": false,
1491
+ "special": false
1492
+ },
1493
+ "186": {
1494
+ "content": "<h2>",
1495
+ "lstrip": false,
1496
+ "normalized": false,
1497
+ "rstrip": false,
1498
+ "single_word": false,
1499
+ "special": false
1500
+ },
1501
+ "187": {
1502
+ "content": "<h3>",
1503
+ "lstrip": false,
1504
+ "normalized": false,
1505
+ "rstrip": false,
1506
+ "single_word": false,
1507
+ "special": false
1508
+ },
1509
+ "188": {
1510
+ "content": "<h4>",
1511
+ "lstrip": false,
1512
+ "normalized": false,
1513
+ "rstrip": false,
1514
+ "single_word": false,
1515
+ "special": false
1516
+ },
1517
+ "189": {
1518
+ "content": "<h5>",
1519
+ "lstrip": false,
1520
+ "normalized": false,
1521
+ "rstrip": false,
1522
+ "single_word": false,
1523
+ "special": false
1524
+ },
1525
+ "190": {
1526
+ "content": "<h6>",
1527
+ "lstrip": false,
1528
+ "normalized": false,
1529
+ "rstrip": false,
1530
+ "single_word": false,
1531
+ "special": false
1532
+ },
1533
+ "191": {
1534
+ "content": "<blockquote>",
1535
+ "lstrip": false,
1536
+ "normalized": false,
1537
+ "rstrip": false,
1538
+ "single_word": false,
1539
+ "special": false
1540
+ },
1541
+ "192": {
1542
+ "content": "</h1>",
1543
+ "lstrip": false,
1544
+ "normalized": false,
1545
+ "rstrip": false,
1546
+ "single_word": false,
1547
+ "special": false
1548
+ },
1549
+ "193": {
1550
+ "content": "</h2>",
1551
+ "lstrip": false,
1552
+ "normalized": false,
1553
+ "rstrip": false,
1554
+ "single_word": false,
1555
+ "special": false
1556
+ },
1557
+ "194": {
1558
+ "content": "</h3>",
1559
+ "lstrip": false,
1560
+ "normalized": false,
1561
+ "rstrip": false,
1562
+ "single_word": false,
1563
+ "special": false
1564
+ },
1565
+ "195": {
1566
+ "content": "</h4>",
1567
+ "lstrip": false,
1568
+ "normalized": false,
1569
+ "rstrip": false,
1570
+ "single_word": false,
1571
+ "special": false
1572
+ },
1573
+ "196": {
1574
+ "content": "</h5>",
1575
+ "lstrip": false,
1576
+ "normalized": false,
1577
+ "rstrip": false,
1578
+ "single_word": false,
1579
+ "special": false
1580
+ },
1581
+ "197": {
1582
+ "content": "</h6>",
1583
+ "lstrip": false,
1584
+ "normalized": false,
1585
+ "rstrip": false,
1586
+ "single_word": false,
1587
+ "special": false
1588
+ },
1589
+ "198": {
1590
+ "content": "</blockquote>",
1591
+ "lstrip": false,
1592
+ "normalized": false,
1593
+ "rstrip": false,
1594
+ "single_word": false,
1595
+ "special": false
1596
+ },
1597
+ "199": {
1598
+ "content": "<strong>",
1599
+ "lstrip": false,
1600
+ "normalized": false,
1601
+ "rstrip": false,
1602
+ "single_word": false,
1603
+ "special": false
1604
+ },
1605
+ "200": {
1606
+ "content": "<em>",
1607
+ "lstrip": false,
1608
+ "normalized": false,
1609
+ "rstrip": false,
1610
+ "single_word": false,
1611
+ "special": false
1612
+ },
1613
+ "201": {
1614
+ "content": "<b>",
1615
+ "lstrip": false,
1616
+ "normalized": false,
1617
+ "rstrip": false,
1618
+ "single_word": false,
1619
+ "special": false
1620
+ },
1621
+ "202": {
1622
+ "content": "<i>",
1623
+ "lstrip": false,
1624
+ "normalized": false,
1625
+ "rstrip": false,
1626
+ "single_word": false,
1627
+ "special": false
1628
+ },
1629
+ "203": {
1630
+ "content": "<u>",
1631
+ "lstrip": false,
1632
+ "normalized": false,
1633
+ "rstrip": false,
1634
+ "single_word": false,
1635
+ "special": false
1636
+ },
1637
+ "204": {
1638
+ "content": "<s>",
1639
+ "lstrip": false,
1640
+ "normalized": false,
1641
+ "rstrip": false,
1642
+ "single_word": false,
1643
+ "special": false
1644
+ },
1645
+ "205": {
1646
+ "content": "<sub>",
1647
+ "lstrip": false,
1648
+ "normalized": false,
1649
+ "rstrip": false,
1650
+ "single_word": false,
1651
+ "special": false
1652
+ },
1653
+ "206": {
1654
+ "content": "<sup>",
1655
+ "lstrip": false,
1656
+ "normalized": false,
1657
+ "rstrip": false,
1658
+ "single_word": false,
1659
+ "special": false
1660
+ },
1661
+ "207": {
1662
+ "content": "<code>",
1663
+ "lstrip": false,
1664
+ "normalized": false,
1665
+ "rstrip": false,
1666
+ "single_word": false,
1667
+ "special": false
1668
+ },
1669
+ "208": {
1670
+ "content": "</strong>",
1671
+ "lstrip": false,
1672
+ "normalized": false,
1673
+ "rstrip": false,
1674
+ "single_word": false,
1675
+ "special": false
1676
+ },
1677
+ "209": {
1678
+ "content": "</em>",
1679
+ "lstrip": false,
1680
+ "normalized": false,
1681
+ "rstrip": false,
1682
+ "single_word": false,
1683
+ "special": false
1684
+ },
1685
+ "210": {
1686
+ "content": "</b>",
1687
+ "lstrip": false,
1688
+ "normalized": false,
1689
+ "rstrip": false,
1690
+ "single_word": false,
1691
+ "special": false
1692
+ },
1693
+ "211": {
1694
+ "content": "</i>",
1695
+ "lstrip": false,
1696
+ "normalized": false,
1697
+ "rstrip": false,
1698
+ "single_word": false,
1699
+ "special": false
1700
+ },
1701
+ "212": {
1702
+ "content": "</u>",
1703
+ "lstrip": false,
1704
+ "normalized": false,
1705
+ "rstrip": false,
1706
+ "single_word": false,
1707
+ "special": false
1708
+ },
1709
+ "213": {
1710
+ "content": "</s>",
1711
+ "lstrip": false,
1712
+ "normalized": false,
1713
+ "rstrip": false,
1714
+ "single_word": false,
1715
+ "special": false
1716
+ },
1717
+ "214": {
1718
+ "content": "</sub>",
1719
+ "lstrip": false,
1720
+ "normalized": false,
1721
+ "rstrip": false,
1722
+ "single_word": false,
1723
+ "special": false
1724
+ },
1725
+ "215": {
1726
+ "content": "</sup>",
1727
+ "lstrip": false,
1728
+ "normalized": false,
1729
+ "rstrip": false,
1730
+ "single_word": false,
1731
+ "special": false
1732
+ },
1733
+ "216": {
1734
+ "content": "</code>",
1735
+ "lstrip": false,
1736
+ "normalized": false,
1737
+ "rstrip": false,
1738
+ "single_word": false,
1739
+ "special": false
1740
+ }
1741
+ },
1742
+ "additional_special_tokens": [
1743
+ "<start_of_turn>",
1744
+ "<end_of_turn>"
1745
+ ],
1746
+ "bos_token": "<bos>",
1747
+ "clean_up_tokenization_spaces": false,
1748
+ "eos_token": "<eos>",
1749
+ "extra_special_tokens": {},
1750
+ "model_max_length": 1000000000000000019884624838656,
1751
+ "pad_token": "<eos>",
1752
+ "padding_side": "right",
1753
+ "sp_model_kwargs": {},
1754
+ "spaces_between_special_tokens": false,
1755
+ "tokenizer_class": "GemmaTokenizer",
1756
+ "unk_token": "<unk>",
1757
+ "use_default_system_prompt": false
1758
+ }
checkpoint-338/trainer_state.json ADDED
@@ -0,0 +1,2400 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 338,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.002962962962962963,
14
+ "grad_norm": 82.5,
15
+ "learning_rate": 0.0,
16
+ "loss": 5.4851,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.005925925925925926,
21
+ "grad_norm": 81.5,
22
+ "learning_rate": 1.0000000000000002e-06,
23
+ "loss": 5.5219,
24
+ "step": 2
25
+ },
26
+ {
27
+ "epoch": 0.008888888888888889,
28
+ "grad_norm": 83.0,
29
+ "learning_rate": 2.0000000000000003e-06,
30
+ "loss": 5.7155,
31
+ "step": 3
32
+ },
33
+ {
34
+ "epoch": 0.011851851851851851,
35
+ "grad_norm": 74.0,
36
+ "learning_rate": 3e-06,
37
+ "loss": 5.5065,
38
+ "step": 4
39
+ },
40
+ {
41
+ "epoch": 0.014814814814814815,
42
+ "grad_norm": 60.5,
43
+ "learning_rate": 4.000000000000001e-06,
44
+ "loss": 4.8696,
45
+ "step": 5
46
+ },
47
+ {
48
+ "epoch": 0.017777777777777778,
49
+ "grad_norm": 48.25,
50
+ "learning_rate": 5e-06,
51
+ "loss": 4.3835,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 0.02074074074074074,
56
+ "grad_norm": 40.5,
57
+ "learning_rate": 6e-06,
58
+ "loss": 4.1231,
59
+ "step": 7
60
+ },
61
+ {
62
+ "epoch": 0.023703703703703703,
63
+ "grad_norm": 50.5,
64
+ "learning_rate": 7e-06,
65
+ "loss": 3.6886,
66
+ "step": 8
67
+ },
68
+ {
69
+ "epoch": 0.02666666666666667,
70
+ "grad_norm": 34.5,
71
+ "learning_rate": 8.000000000000001e-06,
72
+ "loss": 3.3401,
73
+ "step": 9
74
+ },
75
+ {
76
+ "epoch": 0.02962962962962963,
77
+ "grad_norm": 27.125,
78
+ "learning_rate": 9e-06,
79
+ "loss": 3.043,
80
+ "step": 10
81
+ },
82
+ {
83
+ "epoch": 0.03259259259259259,
84
+ "grad_norm": 21.5,
85
+ "learning_rate": 1e-05,
86
+ "loss": 2.7408,
87
+ "step": 11
88
+ },
89
+ {
90
+ "epoch": 0.035555555555555556,
91
+ "grad_norm": 19.875,
92
+ "learning_rate": 1.1000000000000001e-05,
93
+ "loss": 2.5235,
94
+ "step": 12
95
+ },
96
+ {
97
+ "epoch": 0.03851851851851852,
98
+ "grad_norm": 19.0,
99
+ "learning_rate": 1.2e-05,
100
+ "loss": 2.4532,
101
+ "step": 13
102
+ },
103
+ {
104
+ "epoch": 0.04148148148148148,
105
+ "grad_norm": 19.625,
106
+ "learning_rate": 1.3000000000000001e-05,
107
+ "loss": 2.3238,
108
+ "step": 14
109
+ },
110
+ {
111
+ "epoch": 0.044444444444444446,
112
+ "grad_norm": 20.125,
113
+ "learning_rate": 1.4e-05,
114
+ "loss": 2.2851,
115
+ "step": 15
116
+ },
117
+ {
118
+ "epoch": 0.047407407407407405,
119
+ "grad_norm": 17.5,
120
+ "learning_rate": 1.5000000000000002e-05,
121
+ "loss": 2.2676,
122
+ "step": 16
123
+ },
124
+ {
125
+ "epoch": 0.05037037037037037,
126
+ "grad_norm": 16.0,
127
+ "learning_rate": 1.6000000000000003e-05,
128
+ "loss": 2.1502,
129
+ "step": 17
130
+ },
131
+ {
132
+ "epoch": 0.05333333333333334,
133
+ "grad_norm": 14.4375,
134
+ "learning_rate": 1.7e-05,
135
+ "loss": 2.1293,
136
+ "step": 18
137
+ },
138
+ {
139
+ "epoch": 0.056296296296296296,
140
+ "grad_norm": 15.5625,
141
+ "learning_rate": 1.8e-05,
142
+ "loss": 2.0649,
143
+ "step": 19
144
+ },
145
+ {
146
+ "epoch": 0.05925925925925926,
147
+ "grad_norm": 19.0,
148
+ "learning_rate": 1.9e-05,
149
+ "loss": 1.961,
150
+ "step": 20
151
+ },
152
+ {
153
+ "epoch": 0.06222222222222222,
154
+ "grad_norm": 14.375,
155
+ "learning_rate": 2e-05,
156
+ "loss": 1.9368,
157
+ "step": 21
158
+ },
159
+ {
160
+ "epoch": 0.06518518518518518,
161
+ "grad_norm": 15.75,
162
+ "learning_rate": 1.999951200881446e-05,
163
+ "loss": 1.9511,
164
+ "step": 22
165
+ },
166
+ {
167
+ "epoch": 0.06814814814814815,
168
+ "grad_norm": 13.5625,
169
+ "learning_rate": 1.999804808288491e-05,
170
+ "loss": 1.8874,
171
+ "step": 23
172
+ },
173
+ {
174
+ "epoch": 0.07111111111111111,
175
+ "grad_norm": 13.4375,
176
+ "learning_rate": 1.9995608365087945e-05,
177
+ "loss": 1.7678,
178
+ "step": 24
179
+ },
180
+ {
181
+ "epoch": 0.07407407407407407,
182
+ "grad_norm": 14.25,
183
+ "learning_rate": 1.999219309353572e-05,
184
+ "loss": 1.8061,
185
+ "step": 25
186
+ },
187
+ {
188
+ "epoch": 0.07703703703703704,
189
+ "grad_norm": 13.9375,
190
+ "learning_rate": 1.9987802601552717e-05,
191
+ "loss": 1.778,
192
+ "step": 26
193
+ },
194
+ {
195
+ "epoch": 0.08,
196
+ "grad_norm": 11.3125,
197
+ "learning_rate": 1.9982437317643218e-05,
198
+ "loss": 1.5772,
199
+ "step": 27
200
+ },
201
+ {
202
+ "epoch": 0.08296296296296296,
203
+ "grad_norm": 13.3125,
204
+ "learning_rate": 1.9976097765449463e-05,
205
+ "loss": 1.7557,
206
+ "step": 28
207
+ },
208
+ {
209
+ "epoch": 0.08592592592592592,
210
+ "grad_norm": 12.9375,
211
+ "learning_rate": 1.9968784563700586e-05,
212
+ "loss": 1.6816,
213
+ "step": 29
214
+ },
215
+ {
216
+ "epoch": 0.08888888888888889,
217
+ "grad_norm": 12.8125,
218
+ "learning_rate": 1.996049842615217e-05,
219
+ "loss": 1.742,
220
+ "step": 30
221
+ },
222
+ {
223
+ "epoch": 0.09185185185185185,
224
+ "grad_norm": 12.125,
225
+ "learning_rate": 1.9951240161516643e-05,
226
+ "loss": 1.7055,
227
+ "step": 31
228
+ },
229
+ {
230
+ "epoch": 0.09481481481481481,
231
+ "grad_norm": 11.5625,
232
+ "learning_rate": 1.9941010673384307e-05,
233
+ "loss": 1.6907,
234
+ "step": 32
235
+ },
236
+ {
237
+ "epoch": 0.09777777777777778,
238
+ "grad_norm": 11.4375,
239
+ "learning_rate": 1.992981096013517e-05,
240
+ "loss": 1.7213,
241
+ "step": 33
242
+ },
243
+ {
244
+ "epoch": 0.10074074074074074,
245
+ "grad_norm": 12.125,
246
+ "learning_rate": 1.9917642114841505e-05,
247
+ "loss": 1.8787,
248
+ "step": 34
249
+ },
250
+ {
251
+ "epoch": 0.1037037037037037,
252
+ "grad_norm": 11.125,
253
+ "learning_rate": 1.990450532516116e-05,
254
+ "loss": 1.6781,
255
+ "step": 35
256
+ },
257
+ {
258
+ "epoch": 0.10666666666666667,
259
+ "grad_norm": 12.8125,
260
+ "learning_rate": 1.9890401873221642e-05,
261
+ "loss": 1.5917,
262
+ "step": 36
263
+ },
264
+ {
265
+ "epoch": 0.10962962962962963,
266
+ "grad_norm": 12.0625,
267
+ "learning_rate": 1.9875333135495e-05,
268
+ "loss": 1.7399,
269
+ "step": 37
270
+ },
271
+ {
272
+ "epoch": 0.11259259259259259,
273
+ "grad_norm": 11.6875,
274
+ "learning_rate": 1.985930058266348e-05,
275
+ "loss": 1.8376,
276
+ "step": 38
277
+ },
278
+ {
279
+ "epoch": 0.11555555555555555,
280
+ "grad_norm": 10.9375,
281
+ "learning_rate": 1.984230577947597e-05,
282
+ "loss": 1.7302,
283
+ "step": 39
284
+ },
285
+ {
286
+ "epoch": 0.11851851851851852,
287
+ "grad_norm": 10.125,
288
+ "learning_rate": 1.9824350384595295e-05,
289
+ "loss": 1.6194,
290
+ "step": 40
291
+ },
292
+ {
293
+ "epoch": 0.12148148148148148,
294
+ "grad_norm": 10.375,
295
+ "learning_rate": 1.9805436150436352e-05,
296
+ "loss": 1.6648,
297
+ "step": 41
298
+ },
299
+ {
300
+ "epoch": 0.12444444444444444,
301
+ "grad_norm": 12.3125,
302
+ "learning_rate": 1.9785564922995042e-05,
303
+ "loss": 1.7417,
304
+ "step": 42
305
+ },
306
+ {
307
+ "epoch": 0.1274074074074074,
308
+ "grad_norm": 11.625,
309
+ "learning_rate": 1.9764738641668137e-05,
310
+ "loss": 1.6403,
311
+ "step": 43
312
+ },
313
+ {
314
+ "epoch": 0.13037037037037036,
315
+ "grad_norm": 10.625,
316
+ "learning_rate": 1.9742959339063977e-05,
317
+ "loss": 1.6204,
318
+ "step": 44
319
+ },
320
+ {
321
+ "epoch": 0.13333333333333333,
322
+ "grad_norm": 11.75,
323
+ "learning_rate": 1.972022914080411e-05,
324
+ "loss": 1.7575,
325
+ "step": 45
326
+ },
327
+ {
328
+ "epoch": 0.1362962962962963,
329
+ "grad_norm": 12.625,
330
+ "learning_rate": 1.9696550265315805e-05,
331
+ "loss": 1.594,
332
+ "step": 46
333
+ },
334
+ {
335
+ "epoch": 0.13925925925925925,
336
+ "grad_norm": 11.3125,
337
+ "learning_rate": 1.9671925023615572e-05,
338
+ "loss": 1.6259,
339
+ "step": 47
340
+ },
341
+ {
342
+ "epoch": 0.14222222222222222,
343
+ "grad_norm": 11.9375,
344
+ "learning_rate": 1.964635581908359e-05,
345
+ "loss": 1.7574,
346
+ "step": 48
347
+ },
348
+ {
349
+ "epoch": 0.1451851851851852,
350
+ "grad_norm": 10.75,
351
+ "learning_rate": 1.961984514722914e-05,
352
+ "loss": 1.7021,
353
+ "step": 49
354
+ },
355
+ {
356
+ "epoch": 0.14814814814814814,
357
+ "grad_norm": 10.625,
358
+ "learning_rate": 1.9592395595447064e-05,
359
+ "loss": 1.5892,
360
+ "step": 50
361
+ },
362
+ {
363
+ "epoch": 0.1511111111111111,
364
+ "grad_norm": 10.625,
365
+ "learning_rate": 1.9564009842765225e-05,
366
+ "loss": 1.6846,
367
+ "step": 51
368
+ },
369
+ {
370
+ "epoch": 0.15407407407407409,
371
+ "grad_norm": 10.375,
372
+ "learning_rate": 1.9534690659583045e-05,
373
+ "loss": 1.6761,
374
+ "step": 52
375
+ },
376
+ {
377
+ "epoch": 0.15703703703703703,
378
+ "grad_norm": 10.625,
379
+ "learning_rate": 1.9504440907401113e-05,
380
+ "loss": 1.4986,
381
+ "step": 53
382
+ },
383
+ {
384
+ "epoch": 0.16,
385
+ "grad_norm": 11.4375,
386
+ "learning_rate": 1.9473263538541916e-05,
387
+ "loss": 1.6183,
388
+ "step": 54
389
+ },
390
+ {
391
+ "epoch": 0.16296296296296298,
392
+ "grad_norm": 12.3125,
393
+ "learning_rate": 1.944116159586169e-05,
394
+ "loss": 1.7405,
395
+ "step": 55
396
+ },
397
+ {
398
+ "epoch": 0.16592592592592592,
399
+ "grad_norm": 11.75,
400
+ "learning_rate": 1.9408138212453456e-05,
401
+ "loss": 1.608,
402
+ "step": 56
403
+ },
404
+ {
405
+ "epoch": 0.1688888888888889,
406
+ "grad_norm": 10.6875,
407
+ "learning_rate": 1.9374196611341212e-05,
408
+ "loss": 1.6267,
409
+ "step": 57
410
+ },
411
+ {
412
+ "epoch": 0.17185185185185184,
413
+ "grad_norm": 12.3125,
414
+ "learning_rate": 1.933934010516539e-05,
415
+ "loss": 1.7335,
416
+ "step": 58
417
+ },
418
+ {
419
+ "epoch": 0.1748148148148148,
420
+ "grad_norm": 11.125,
421
+ "learning_rate": 1.9303572095859545e-05,
422
+ "loss": 1.7112,
423
+ "step": 59
424
+ },
425
+ {
426
+ "epoch": 0.17777777777777778,
427
+ "grad_norm": 11.75,
428
+ "learning_rate": 1.9266896074318335e-05,
429
+ "loss": 1.8106,
430
+ "step": 60
431
+ },
432
+ {
433
+ "epoch": 0.18074074074074073,
434
+ "grad_norm": 10.875,
435
+ "learning_rate": 1.9229315620056805e-05,
436
+ "loss": 1.5596,
437
+ "step": 61
438
+ },
439
+ {
440
+ "epoch": 0.1837037037037037,
441
+ "grad_norm": 14.0625,
442
+ "learning_rate": 1.9190834400861035e-05,
443
+ "loss": 1.5379,
444
+ "step": 62
445
+ },
446
+ {
447
+ "epoch": 0.18666666666666668,
448
+ "grad_norm": 10.875,
449
+ "learning_rate": 1.9151456172430186e-05,
450
+ "loss": 1.5921,
451
+ "step": 63
452
+ },
453
+ {
454
+ "epoch": 0.18962962962962962,
455
+ "grad_norm": 10.75,
456
+ "learning_rate": 1.9111184778009934e-05,
457
+ "loss": 1.6252,
458
+ "step": 64
459
+ },
460
+ {
461
+ "epoch": 0.1925925925925926,
462
+ "grad_norm": 11.8125,
463
+ "learning_rate": 1.9070024148017375e-05,
464
+ "loss": 1.7589,
465
+ "step": 65
466
+ },
467
+ {
468
+ "epoch": 0.19555555555555557,
469
+ "grad_norm": 10.4375,
470
+ "learning_rate": 1.9027978299657436e-05,
471
+ "loss": 1.4877,
472
+ "step": 66
473
+ },
474
+ {
475
+ "epoch": 0.1985185185185185,
476
+ "grad_norm": 10.0625,
477
+ "learning_rate": 1.89850513365308e-05,
478
+ "loss": 1.6578,
479
+ "step": 67
480
+ },
481
+ {
482
+ "epoch": 0.20148148148148148,
483
+ "grad_norm": 11.5,
484
+ "learning_rate": 1.8941247448233386e-05,
485
+ "loss": 1.5563,
486
+ "step": 68
487
+ },
488
+ {
489
+ "epoch": 0.20444444444444446,
490
+ "grad_norm": 12.25,
491
+ "learning_rate": 1.8896570909947477e-05,
492
+ "loss": 1.6964,
493
+ "step": 69
494
+ },
495
+ {
496
+ "epoch": 0.2074074074074074,
497
+ "grad_norm": 10.5625,
498
+ "learning_rate": 1.885102608202444e-05,
499
+ "loss": 1.5913,
500
+ "step": 70
501
+ },
502
+ {
503
+ "epoch": 0.21037037037037037,
504
+ "grad_norm": 10.625,
505
+ "learning_rate": 1.88046174095592e-05,
506
+ "loss": 1.5556,
507
+ "step": 71
508
+ },
509
+ {
510
+ "epoch": 0.21333333333333335,
511
+ "grad_norm": 10.0625,
512
+ "learning_rate": 1.875734942195637e-05,
513
+ "loss": 1.5116,
514
+ "step": 72
515
+ },
516
+ {
517
+ "epoch": 0.2162962962962963,
518
+ "grad_norm": 9.6875,
519
+ "learning_rate": 1.8709226732488216e-05,
520
+ "loss": 1.5118,
521
+ "step": 73
522
+ },
523
+ {
524
+ "epoch": 0.21925925925925926,
525
+ "grad_norm": 10.4375,
526
+ "learning_rate": 1.866025403784439e-05,
527
+ "loss": 1.6544,
528
+ "step": 74
529
+ },
530
+ {
531
+ "epoch": 0.2222222222222222,
532
+ "grad_norm": 10.875,
533
+ "learning_rate": 1.8610436117673557e-05,
534
+ "loss": 1.5481,
535
+ "step": 75
536
+ },
537
+ {
538
+ "epoch": 0.22518518518518518,
539
+ "grad_norm": 10.75,
540
+ "learning_rate": 1.8559777834116906e-05,
541
+ "loss": 1.5817,
542
+ "step": 76
543
+ },
544
+ {
545
+ "epoch": 0.22814814814814816,
546
+ "grad_norm": 10.5,
547
+ "learning_rate": 1.8508284131333604e-05,
548
+ "loss": 1.5016,
549
+ "step": 77
550
+ },
551
+ {
552
+ "epoch": 0.2311111111111111,
553
+ "grad_norm": 10.8125,
554
+ "learning_rate": 1.845596003501826e-05,
555
+ "loss": 1.5924,
556
+ "step": 78
557
+ },
558
+ {
559
+ "epoch": 0.23407407407407407,
560
+ "grad_norm": 10.25,
561
+ "learning_rate": 1.8402810651910444e-05,
562
+ "loss": 1.63,
563
+ "step": 79
564
+ },
565
+ {
566
+ "epoch": 0.23703703703703705,
567
+ "grad_norm": 10.5625,
568
+ "learning_rate": 1.8348841169296247e-05,
569
+ "loss": 1.5974,
570
+ "step": 80
571
+ },
572
+ {
573
+ "epoch": 0.24,
574
+ "grad_norm": 9.875,
575
+ "learning_rate": 1.829405685450202e-05,
576
+ "loss": 1.5722,
577
+ "step": 81
578
+ },
579
+ {
580
+ "epoch": 0.24296296296296296,
581
+ "grad_norm": 10.5,
582
+ "learning_rate": 1.823846305438032e-05,
583
+ "loss": 1.4561,
584
+ "step": 82
585
+ },
586
+ {
587
+ "epoch": 0.24592592592592594,
588
+ "grad_norm": 9.5625,
589
+ "learning_rate": 1.8182065194788024e-05,
590
+ "loss": 1.4567,
591
+ "step": 83
592
+ },
593
+ {
594
+ "epoch": 0.24888888888888888,
595
+ "grad_norm": 9.875,
596
+ "learning_rate": 1.8124868780056814e-05,
597
+ "loss": 1.613,
598
+ "step": 84
599
+ },
600
+ {
601
+ "epoch": 0.2518518518518518,
602
+ "grad_norm": 9.75,
603
+ "learning_rate": 1.8066879392455932e-05,
604
+ "loss": 1.609,
605
+ "step": 85
606
+ },
607
+ {
608
+ "epoch": 0.2548148148148148,
609
+ "grad_norm": 10.1875,
610
+ "learning_rate": 1.800810269164738e-05,
611
+ "loss": 1.6807,
612
+ "step": 86
613
+ },
614
+ {
615
+ "epoch": 0.2577777777777778,
616
+ "grad_norm": 10.3125,
617
+ "learning_rate": 1.7948544414133534e-05,
618
+ "loss": 1.6024,
619
+ "step": 87
620
+ },
621
+ {
622
+ "epoch": 0.2607407407407407,
623
+ "grad_norm": 9.9375,
624
+ "learning_rate": 1.7888210372697292e-05,
625
+ "loss": 1.4733,
626
+ "step": 88
627
+ },
628
+ {
629
+ "epoch": 0.2637037037037037,
630
+ "grad_norm": 10.4375,
631
+ "learning_rate": 1.782710645583473e-05,
632
+ "loss": 1.5994,
633
+ "step": 89
634
+ },
635
+ {
636
+ "epoch": 0.26666666666666666,
637
+ "grad_norm": 13.625,
638
+ "learning_rate": 1.7765238627180424e-05,
639
+ "loss": 1.3751,
640
+ "step": 90
641
+ },
642
+ {
643
+ "epoch": 0.2696296296296296,
644
+ "grad_norm": 10.875,
645
+ "learning_rate": 1.7702612924925377e-05,
646
+ "loss": 1.436,
647
+ "step": 91
648
+ },
649
+ {
650
+ "epoch": 0.2725925925925926,
651
+ "grad_norm": 9.625,
652
+ "learning_rate": 1.7639235461227727e-05,
653
+ "loss": 1.5887,
654
+ "step": 92
655
+ },
656
+ {
657
+ "epoch": 0.27555555555555555,
658
+ "grad_norm": 9.6875,
659
+ "learning_rate": 1.7575112421616203e-05,
660
+ "loss": 1.5977,
661
+ "step": 93
662
+ },
663
+ {
664
+ "epoch": 0.2785185185185185,
665
+ "grad_norm": 10.25,
666
+ "learning_rate": 1.751025006438643e-05,
667
+ "loss": 1.518,
668
+ "step": 94
669
+ },
670
+ {
671
+ "epoch": 0.2814814814814815,
672
+ "grad_norm": 10.8125,
673
+ "learning_rate": 1.7444654719990128e-05,
674
+ "loss": 1.6543,
675
+ "step": 95
676
+ },
677
+ {
678
+ "epoch": 0.28444444444444444,
679
+ "grad_norm": 9.9375,
680
+ "learning_rate": 1.7378332790417275e-05,
681
+ "loss": 1.5368,
682
+ "step": 96
683
+ },
684
+ {
685
+ "epoch": 0.2874074074074074,
686
+ "grad_norm": 10.0,
687
+ "learning_rate": 1.7311290748571273e-05,
688
+ "loss": 1.5349,
689
+ "step": 97
690
+ },
691
+ {
692
+ "epoch": 0.2903703703703704,
693
+ "grad_norm": 9.875,
694
+ "learning_rate": 1.7243535137637227e-05,
695
+ "loss": 1.6126,
696
+ "step": 98
697
+ },
698
+ {
699
+ "epoch": 0.29333333333333333,
700
+ "grad_norm": 9.8125,
701
+ "learning_rate": 1.717507257044331e-05,
702
+ "loss": 1.5037,
703
+ "step": 99
704
+ },
705
+ {
706
+ "epoch": 0.2962962962962963,
707
+ "grad_norm": 9.3125,
708
+ "learning_rate": 1.7105909728815398e-05,
709
+ "loss": 1.5038,
710
+ "step": 100
711
+ },
712
+ {
713
+ "epoch": 0.2992592592592593,
714
+ "grad_norm": 9.0,
715
+ "learning_rate": 1.7036053362924896e-05,
716
+ "loss": 1.3771,
717
+ "step": 101
718
+ },
719
+ {
720
+ "epoch": 0.3022222222222222,
721
+ "grad_norm": 9.375,
722
+ "learning_rate": 1.6965510290629973e-05,
723
+ "loss": 1.5432,
724
+ "step": 102
725
+ },
726
+ {
727
+ "epoch": 0.30518518518518517,
728
+ "grad_norm": 9.375,
729
+ "learning_rate": 1.689428739681012e-05,
730
+ "loss": 1.4479,
731
+ "step": 103
732
+ },
733
+ {
734
+ "epoch": 0.30814814814814817,
735
+ "grad_norm": 11.1875,
736
+ "learning_rate": 1.682239163269422e-05,
737
+ "loss": 1.5956,
738
+ "step": 104
739
+ },
740
+ {
741
+ "epoch": 0.3111111111111111,
742
+ "grad_norm": 10.5,
743
+ "learning_rate": 1.6749830015182106e-05,
744
+ "loss": 1.4063,
745
+ "step": 105
746
+ },
747
+ {
748
+ "epoch": 0.31407407407407406,
749
+ "grad_norm": 10.0625,
750
+ "learning_rate": 1.667660962615973e-05,
751
+ "loss": 1.6329,
752
+ "step": 106
753
+ },
754
+ {
755
+ "epoch": 0.31703703703703706,
756
+ "grad_norm": 11.3125,
757
+ "learning_rate": 1.6602737611807975e-05,
758
+ "loss": 1.5915,
759
+ "step": 107
760
+ },
761
+ {
762
+ "epoch": 0.32,
763
+ "grad_norm": 9.375,
764
+ "learning_rate": 1.6528221181905217e-05,
765
+ "loss": 1.3679,
766
+ "step": 108
767
+ },
768
+ {
769
+ "epoch": 0.32296296296296295,
770
+ "grad_norm": 9.1875,
771
+ "learning_rate": 1.6453067609123656e-05,
772
+ "loss": 1.4718,
773
+ "step": 109
774
+ },
775
+ {
776
+ "epoch": 0.32592592592592595,
777
+ "grad_norm": 9.6875,
778
+ "learning_rate": 1.6377284228319496e-05,
779
+ "loss": 1.4929,
780
+ "step": 110
781
+ },
782
+ {
783
+ "epoch": 0.3288888888888889,
784
+ "grad_norm": 9.9375,
785
+ "learning_rate": 1.6300878435817115e-05,
786
+ "loss": 1.4887,
787
+ "step": 111
788
+ },
789
+ {
790
+ "epoch": 0.33185185185185184,
791
+ "grad_norm": 10.1875,
792
+ "learning_rate": 1.622385768868716e-05,
793
+ "loss": 1.5634,
794
+ "step": 112
795
+ },
796
+ {
797
+ "epoch": 0.3348148148148148,
798
+ "grad_norm": 9.75,
799
+ "learning_rate": 1.6146229504018777e-05,
800
+ "loss": 1.4839,
801
+ "step": 113
802
+ },
803
+ {
804
+ "epoch": 0.3377777777777778,
805
+ "grad_norm": 10.9375,
806
+ "learning_rate": 1.6068001458185934e-05,
807
+ "loss": 1.4955,
808
+ "step": 114
809
+ },
810
+ {
811
+ "epoch": 0.34074074074074073,
812
+ "grad_norm": 9.5625,
813
+ "learning_rate": 1.5989181186108003e-05,
814
+ "loss": 1.4464,
815
+ "step": 115
816
+ },
817
+ {
818
+ "epoch": 0.3437037037037037,
819
+ "grad_norm": 9.8125,
820
+ "learning_rate": 1.5909776380504583e-05,
821
+ "loss": 1.4568,
822
+ "step": 116
823
+ },
824
+ {
825
+ "epoch": 0.3466666666666667,
826
+ "grad_norm": 10.4375,
827
+ "learning_rate": 1.5829794791144723e-05,
828
+ "loss": 1.5491,
829
+ "step": 117
830
+ },
831
+ {
832
+ "epoch": 0.3496296296296296,
833
+ "grad_norm": 10.9375,
834
+ "learning_rate": 1.5749244224090537e-05,
835
+ "loss": 1.4774,
836
+ "step": 118
837
+ },
838
+ {
839
+ "epoch": 0.35259259259259257,
840
+ "grad_norm": 10.875,
841
+ "learning_rate": 1.566813254093538e-05,
842
+ "loss": 1.5642,
843
+ "step": 119
844
+ },
845
+ {
846
+ "epoch": 0.35555555555555557,
847
+ "grad_norm": 10.4375,
848
+ "learning_rate": 1.5586467658036526e-05,
849
+ "loss": 1.5229,
850
+ "step": 120
851
+ },
852
+ {
853
+ "epoch": 0.3585185185185185,
854
+ "grad_norm": 10.1875,
855
+ "learning_rate": 1.5504257545742585e-05,
856
+ "loss": 1.5067,
857
+ "step": 121
858
+ },
859
+ {
860
+ "epoch": 0.36148148148148146,
861
+ "grad_norm": 9.6875,
862
+ "learning_rate": 1.5421510227615587e-05,
863
+ "loss": 1.4074,
864
+ "step": 122
865
+ },
866
+ {
867
+ "epoch": 0.36444444444444446,
868
+ "grad_norm": 11.3125,
869
+ "learning_rate": 1.533823377964791e-05,
870
+ "loss": 1.4726,
871
+ "step": 123
872
+ },
873
+ {
874
+ "epoch": 0.3674074074074074,
875
+ "grad_norm": 12.75,
876
+ "learning_rate": 1.5254436329474062e-05,
877
+ "loss": 1.351,
878
+ "step": 124
879
+ },
880
+ {
881
+ "epoch": 0.37037037037037035,
882
+ "grad_norm": 10.6875,
883
+ "learning_rate": 1.517012605557746e-05,
884
+ "loss": 1.5042,
885
+ "step": 125
886
+ },
887
+ {
888
+ "epoch": 0.37333333333333335,
889
+ "grad_norm": 9.8125,
890
+ "learning_rate": 1.5085311186492206e-05,
891
+ "loss": 1.5669,
892
+ "step": 126
893
+ },
894
+ {
895
+ "epoch": 0.3762962962962963,
896
+ "grad_norm": 9.75,
897
+ "learning_rate": 1.5000000000000002e-05,
898
+ "loss": 1.4575,
899
+ "step": 127
900
+ },
901
+ {
902
+ "epoch": 0.37925925925925924,
903
+ "grad_norm": 9.9375,
904
+ "learning_rate": 1.4914200822322255e-05,
905
+ "loss": 1.5091,
906
+ "step": 128
907
+ },
908
+ {
909
+ "epoch": 0.38222222222222224,
910
+ "grad_norm": 12.5,
911
+ "learning_rate": 1.482792202730745e-05,
912
+ "loss": 1.6976,
913
+ "step": 129
914
+ },
915
+ {
916
+ "epoch": 0.3851851851851852,
917
+ "grad_norm": 11.625,
918
+ "learning_rate": 1.4741172035613884e-05,
919
+ "loss": 1.5555,
920
+ "step": 130
921
+ },
922
+ {
923
+ "epoch": 0.38814814814814813,
924
+ "grad_norm": 10.3125,
925
+ "learning_rate": 1.4653959313887813e-05,
926
+ "loss": 1.7106,
927
+ "step": 131
928
+ },
929
+ {
930
+ "epoch": 0.39111111111111113,
931
+ "grad_norm": 10.3125,
932
+ "learning_rate": 1.4566292373937133e-05,
933
+ "loss": 1.4944,
934
+ "step": 132
935
+ },
936
+ {
937
+ "epoch": 0.3940740740740741,
938
+ "grad_norm": 9.3125,
939
+ "learning_rate": 1.4478179771900634e-05,
940
+ "loss": 1.4851,
941
+ "step": 133
942
+ },
943
+ {
944
+ "epoch": 0.397037037037037,
945
+ "grad_norm": 9.0,
946
+ "learning_rate": 1.4389630107412942e-05,
947
+ "loss": 1.5191,
948
+ "step": 134
949
+ },
950
+ {
951
+ "epoch": 0.4,
952
+ "grad_norm": 10.625,
953
+ "learning_rate": 1.4300652022765207e-05,
954
+ "loss": 1.3619,
955
+ "step": 135
956
+ },
957
+ {
958
+ "epoch": 0.40296296296296297,
959
+ "grad_norm": 10.4375,
960
+ "learning_rate": 1.4211254202061632e-05,
961
+ "loss": 1.457,
962
+ "step": 136
963
+ },
964
+ {
965
+ "epoch": 0.4059259259259259,
966
+ "grad_norm": 9.3125,
967
+ "learning_rate": 1.4121445370371922e-05,
968
+ "loss": 1.4453,
969
+ "step": 137
970
+ },
971
+ {
972
+ "epoch": 0.4088888888888889,
973
+ "grad_norm": 9.375,
974
+ "learning_rate": 1.4031234292879726e-05,
975
+ "loss": 1.4961,
976
+ "step": 138
977
+ },
978
+ {
979
+ "epoch": 0.41185185185185186,
980
+ "grad_norm": 9.625,
981
+ "learning_rate": 1.394062977402717e-05,
982
+ "loss": 1.4724,
983
+ "step": 139
984
+ },
985
+ {
986
+ "epoch": 0.4148148148148148,
987
+ "grad_norm": 10.3125,
988
+ "learning_rate": 1.3849640656655572e-05,
989
+ "loss": 1.4337,
990
+ "step": 140
991
+ },
992
+ {
993
+ "epoch": 0.4177777777777778,
994
+ "grad_norm": 11.1875,
995
+ "learning_rate": 1.3758275821142382e-05,
996
+ "loss": 1.4608,
997
+ "step": 141
998
+ },
999
+ {
1000
+ "epoch": 0.42074074074074075,
1001
+ "grad_norm": 10.0,
1002
+ "learning_rate": 1.3666544184534483e-05,
1003
+ "loss": 1.4542,
1004
+ "step": 142
1005
+ },
1006
+ {
1007
+ "epoch": 0.4237037037037037,
1008
+ "grad_norm": 10.6875,
1009
+ "learning_rate": 1.3574454699677893e-05,
1010
+ "loss": 1.5512,
1011
+ "step": 143
1012
+ },
1013
+ {
1014
+ "epoch": 0.4266666666666667,
1015
+ "grad_norm": 10.3125,
1016
+ "learning_rate": 1.348201635434399e-05,
1017
+ "loss": 1.56,
1018
+ "step": 144
1019
+ },
1020
+ {
1021
+ "epoch": 0.42962962962962964,
1022
+ "grad_norm": 9.625,
1023
+ "learning_rate": 1.3389238170352318e-05,
1024
+ "loss": 1.3596,
1025
+ "step": 145
1026
+ },
1027
+ {
1028
+ "epoch": 0.4325925925925926,
1029
+ "grad_norm": 9.5,
1030
+ "learning_rate": 1.329612920269008e-05,
1031
+ "loss": 1.4438,
1032
+ "step": 146
1033
+ },
1034
+ {
1035
+ "epoch": 0.43555555555555553,
1036
+ "grad_norm": 10.1875,
1037
+ "learning_rate": 1.3202698538628376e-05,
1038
+ "loss": 1.4873,
1039
+ "step": 147
1040
+ },
1041
+ {
1042
+ "epoch": 0.43851851851851853,
1043
+ "grad_norm": 9.8125,
1044
+ "learning_rate": 1.3108955296835313e-05,
1045
+ "loss": 1.4859,
1046
+ "step": 148
1047
+ },
1048
+ {
1049
+ "epoch": 0.4414814814814815,
1050
+ "grad_norm": 9.75,
1051
+ "learning_rate": 1.3014908626486032e-05,
1052
+ "loss": 1.4505,
1053
+ "step": 149
1054
+ },
1055
+ {
1056
+ "epoch": 0.4444444444444444,
1057
+ "grad_norm": 9.9375,
1058
+ "learning_rate": 1.292056770636976e-05,
1059
+ "loss": 1.3619,
1060
+ "step": 150
1061
+ },
1062
+ {
1063
+ "epoch": 0.4474074074074074,
1064
+ "grad_norm": 10.25,
1065
+ "learning_rate": 1.282594174399399e-05,
1066
+ "loss": 1.4956,
1067
+ "step": 151
1068
+ },
1069
+ {
1070
+ "epoch": 0.45037037037037037,
1071
+ "grad_norm": 9.6875,
1072
+ "learning_rate": 1.2731039974685833e-05,
1073
+ "loss": 1.4486,
1074
+ "step": 152
1075
+ },
1076
+ {
1077
+ "epoch": 0.4533333333333333,
1078
+ "grad_norm": 9.5625,
1079
+ "learning_rate": 1.2635871660690677e-05,
1080
+ "loss": 1.377,
1081
+ "step": 153
1082
+ },
1083
+ {
1084
+ "epoch": 0.4562962962962963,
1085
+ "grad_norm": 9.4375,
1086
+ "learning_rate": 1.2540446090268193e-05,
1087
+ "loss": 1.4245,
1088
+ "step": 154
1089
+ },
1090
+ {
1091
+ "epoch": 0.45925925925925926,
1092
+ "grad_norm": 9.375,
1093
+ "learning_rate": 1.2444772576785828e-05,
1094
+ "loss": 1.407,
1095
+ "step": 155
1096
+ },
1097
+ {
1098
+ "epoch": 0.4622222222222222,
1099
+ "grad_norm": 9.5625,
1100
+ "learning_rate": 1.234886045780984e-05,
1101
+ "loss": 1.4341,
1102
+ "step": 156
1103
+ },
1104
+ {
1105
+ "epoch": 0.4651851851851852,
1106
+ "grad_norm": 9.3125,
1107
+ "learning_rate": 1.225271909419395e-05,
1108
+ "loss": 1.5296,
1109
+ "step": 157
1110
+ },
1111
+ {
1112
+ "epoch": 0.46814814814814815,
1113
+ "grad_norm": 9.0625,
1114
+ "learning_rate": 1.2156357869165771e-05,
1115
+ "loss": 1.4201,
1116
+ "step": 158
1117
+ },
1118
+ {
1119
+ "epoch": 0.4711111111111111,
1120
+ "grad_norm": 9.0625,
1121
+ "learning_rate": 1.2059786187410984e-05,
1122
+ "loss": 1.458,
1123
+ "step": 159
1124
+ },
1125
+ {
1126
+ "epoch": 0.4740740740740741,
1127
+ "grad_norm": 9.8125,
1128
+ "learning_rate": 1.1963013474155487e-05,
1129
+ "loss": 1.5306,
1130
+ "step": 160
1131
+ },
1132
+ {
1133
+ "epoch": 0.47703703703703704,
1134
+ "grad_norm": 9.4375,
1135
+ "learning_rate": 1.186604917424549e-05,
1136
+ "loss": 1.437,
1137
+ "step": 161
1138
+ },
1139
+ {
1140
+ "epoch": 0.48,
1141
+ "grad_norm": 9.5,
1142
+ "learning_rate": 1.176890275122573e-05,
1143
+ "loss": 1.4678,
1144
+ "step": 162
1145
+ },
1146
+ {
1147
+ "epoch": 0.482962962962963,
1148
+ "grad_norm": 9.0625,
1149
+ "learning_rate": 1.1671583686415833e-05,
1150
+ "loss": 1.2917,
1151
+ "step": 163
1152
+ },
1153
+ {
1154
+ "epoch": 0.48592592592592593,
1155
+ "grad_norm": 9.25,
1156
+ "learning_rate": 1.1574101477984966e-05,
1157
+ "loss": 1.4942,
1158
+ "step": 164
1159
+ },
1160
+ {
1161
+ "epoch": 0.4888888888888889,
1162
+ "grad_norm": 9.5625,
1163
+ "learning_rate": 1.1476465640024814e-05,
1164
+ "loss": 1.458,
1165
+ "step": 165
1166
+ },
1167
+ {
1168
+ "epoch": 0.4918518518518519,
1169
+ "grad_norm": 10.0625,
1170
+ "learning_rate": 1.1378685701621047e-05,
1171
+ "loss": 1.4213,
1172
+ "step": 166
1173
+ },
1174
+ {
1175
+ "epoch": 0.4948148148148148,
1176
+ "grad_norm": 8.625,
1177
+ "learning_rate": 1.1280771205923269e-05,
1178
+ "loss": 1.2903,
1179
+ "step": 167
1180
+ },
1181
+ {
1182
+ "epoch": 0.49777777777777776,
1183
+ "grad_norm": 9.125,
1184
+ "learning_rate": 1.1182731709213658e-05,
1185
+ "loss": 1.3953,
1186
+ "step": 168
1187
+ },
1188
+ {
1189
+ "epoch": 0.5007407407407407,
1190
+ "grad_norm": 9.5,
1191
+ "learning_rate": 1.1084576779974257e-05,
1192
+ "loss": 1.4805,
1193
+ "step": 169
1194
+ },
1195
+ {
1196
+ "epoch": 0.5037037037037037,
1197
+ "grad_norm": 9.75,
1198
+ "learning_rate": 1.0986315997953118e-05,
1199
+ "loss": 1.4822,
1200
+ "step": 170
1201
+ },
1202
+ {
1203
+ "epoch": 0.5066666666666667,
1204
+ "grad_norm": 9.75,
1205
+ "learning_rate": 1.0887958953229349e-05,
1206
+ "loss": 1.4369,
1207
+ "step": 171
1208
+ },
1209
+ {
1210
+ "epoch": 0.5096296296296297,
1211
+ "grad_norm": 9.375,
1212
+ "learning_rate": 1.078951524527712e-05,
1213
+ "loss": 1.3851,
1214
+ "step": 172
1215
+ },
1216
+ {
1217
+ "epoch": 0.5125925925925926,
1218
+ "grad_norm": 10.375,
1219
+ "learning_rate": 1.069099448202878e-05,
1220
+ "loss": 1.6454,
1221
+ "step": 173
1222
+ },
1223
+ {
1224
+ "epoch": 0.5155555555555555,
1225
+ "grad_norm": 9.1875,
1226
+ "learning_rate": 1.0592406278937143e-05,
1227
+ "loss": 1.4134,
1228
+ "step": 174
1229
+ },
1230
+ {
1231
+ "epoch": 0.5185185185185185,
1232
+ "grad_norm": 9.625,
1233
+ "learning_rate": 1.049376025803703e-05,
1234
+ "loss": 1.3613,
1235
+ "step": 175
1236
+ },
1237
+ {
1238
+ "epoch": 0.5214814814814814,
1239
+ "grad_norm": 9.5625,
1240
+ "learning_rate": 1.039506604700618e-05,
1241
+ "loss": 1.3413,
1242
+ "step": 176
1243
+ },
1244
+ {
1245
+ "epoch": 0.5244444444444445,
1246
+ "grad_norm": 8.875,
1247
+ "learning_rate": 1.0296333278225599e-05,
1248
+ "loss": 1.3768,
1249
+ "step": 177
1250
+ },
1251
+ {
1252
+ "epoch": 0.5274074074074074,
1253
+ "grad_norm": 9.875,
1254
+ "learning_rate": 1.0197571587839466e-05,
1255
+ "loss": 1.4624,
1256
+ "step": 178
1257
+ },
1258
+ {
1259
+ "epoch": 0.5303703703703704,
1260
+ "grad_norm": 9.5,
1261
+ "learning_rate": 1.0098790614814658e-05,
1262
+ "loss": 1.3982,
1263
+ "step": 179
1264
+ },
1265
+ {
1266
+ "epoch": 0.5333333333333333,
1267
+ "grad_norm": 9.0,
1268
+ "learning_rate": 1e-05,
1269
+ "loss": 1.4345,
1270
+ "step": 180
1271
+ },
1272
+ {
1273
+ "epoch": 0.5362962962962963,
1274
+ "grad_norm": 9.4375,
1275
+ "learning_rate": 9.901209385185345e-06,
1276
+ "loss": 1.4967,
1277
+ "step": 181
1278
+ },
1279
+ {
1280
+ "epoch": 0.5392592592592592,
1281
+ "grad_norm": 10.0625,
1282
+ "learning_rate": 9.802428412160538e-06,
1283
+ "loss": 1.5533,
1284
+ "step": 182
1285
+ },
1286
+ {
1287
+ "epoch": 0.5422222222222223,
1288
+ "grad_norm": 10.375,
1289
+ "learning_rate": 9.703666721774403e-06,
1290
+ "loss": 1.4622,
1291
+ "step": 183
1292
+ },
1293
+ {
1294
+ "epoch": 0.5451851851851852,
1295
+ "grad_norm": 9.4375,
1296
+ "learning_rate": 9.604933952993822e-06,
1297
+ "loss": 1.4235,
1298
+ "step": 184
1299
+ },
1300
+ {
1301
+ "epoch": 0.5481481481481482,
1302
+ "grad_norm": 9.5,
1303
+ "learning_rate": 9.506239741962971e-06,
1304
+ "loss": 1.4563,
1305
+ "step": 185
1306
+ },
1307
+ {
1308
+ "epoch": 0.5511111111111111,
1309
+ "grad_norm": 9.6875,
1310
+ "learning_rate": 9.407593721062858e-06,
1311
+ "loss": 1.5149,
1312
+ "step": 186
1313
+ },
1314
+ {
1315
+ "epoch": 0.554074074074074,
1316
+ "grad_norm": 9.25,
1317
+ "learning_rate": 9.309005517971222e-06,
1318
+ "loss": 1.4514,
1319
+ "step": 187
1320
+ },
1321
+ {
1322
+ "epoch": 0.557037037037037,
1323
+ "grad_norm": 9.8125,
1324
+ "learning_rate": 9.210484754722882e-06,
1325
+ "loss": 1.4964,
1326
+ "step": 188
1327
+ },
1328
+ {
1329
+ "epoch": 0.56,
1330
+ "grad_norm": 9.375,
1331
+ "learning_rate": 9.112041046770653e-06,
1332
+ "loss": 1.3697,
1333
+ "step": 189
1334
+ },
1335
+ {
1336
+ "epoch": 0.562962962962963,
1337
+ "grad_norm": 10.25,
1338
+ "learning_rate": 9.013684002046883e-06,
1339
+ "loss": 1.2907,
1340
+ "step": 190
1341
+ },
1342
+ {
1343
+ "epoch": 0.5659259259259259,
1344
+ "grad_norm": 9.625,
1345
+ "learning_rate": 8.915423220025747e-06,
1346
+ "loss": 1.3109,
1347
+ "step": 191
1348
+ },
1349
+ {
1350
+ "epoch": 0.5688888888888889,
1351
+ "grad_norm": 9.25,
1352
+ "learning_rate": 8.817268290786343e-06,
1353
+ "loss": 1.4432,
1354
+ "step": 192
1355
+ },
1356
+ {
1357
+ "epoch": 0.5718518518518518,
1358
+ "grad_norm": 9.6875,
1359
+ "learning_rate": 8.719228794076733e-06,
1360
+ "loss": 1.4068,
1361
+ "step": 193
1362
+ },
1363
+ {
1364
+ "epoch": 0.5748148148148148,
1365
+ "grad_norm": 10.75,
1366
+ "learning_rate": 8.621314298378958e-06,
1367
+ "loss": 1.4999,
1368
+ "step": 194
1369
+ },
1370
+ {
1371
+ "epoch": 0.5777777777777777,
1372
+ "grad_norm": 9.875,
1373
+ "learning_rate": 8.52353435997519e-06,
1374
+ "loss": 1.5483,
1375
+ "step": 195
1376
+ },
1377
+ {
1378
+ "epoch": 0.5807407407407408,
1379
+ "grad_norm": 9.25,
1380
+ "learning_rate": 8.425898522015038e-06,
1381
+ "loss": 1.3818,
1382
+ "step": 196
1383
+ },
1384
+ {
1385
+ "epoch": 0.5837037037037037,
1386
+ "grad_norm": 9.8125,
1387
+ "learning_rate": 8.328416313584169e-06,
1388
+ "loss": 1.4961,
1389
+ "step": 197
1390
+ },
1391
+ {
1392
+ "epoch": 0.5866666666666667,
1393
+ "grad_norm": 9.8125,
1394
+ "learning_rate": 8.231097248774273e-06,
1395
+ "loss": 1.4898,
1396
+ "step": 198
1397
+ },
1398
+ {
1399
+ "epoch": 0.5896296296296296,
1400
+ "grad_norm": 9.4375,
1401
+ "learning_rate": 8.133950825754511e-06,
1402
+ "loss": 1.3789,
1403
+ "step": 199
1404
+ },
1405
+ {
1406
+ "epoch": 0.5925925925925926,
1407
+ "grad_norm": 9.5625,
1408
+ "learning_rate": 8.036986525844516e-06,
1409
+ "loss": 1.3425,
1410
+ "step": 200
1411
+ },
1412
+ {
1413
+ "epoch": 0.5955555555555555,
1414
+ "grad_norm": 10.4375,
1415
+ "learning_rate": 7.940213812589018e-06,
1416
+ "loss": 1.6482,
1417
+ "step": 201
1418
+ },
1419
+ {
1420
+ "epoch": 0.5985185185185186,
1421
+ "grad_norm": 9.875,
1422
+ "learning_rate": 7.843642130834232e-06,
1423
+ "loss": 1.4535,
1424
+ "step": 202
1425
+ },
1426
+ {
1427
+ "epoch": 0.6014814814814815,
1428
+ "grad_norm": 9.5625,
1429
+ "learning_rate": 7.747280905806051e-06,
1430
+ "loss": 1.3676,
1431
+ "step": 203
1432
+ },
1433
+ {
1434
+ "epoch": 0.6044444444444445,
1435
+ "grad_norm": 9.6875,
1436
+ "learning_rate": 7.651139542190164e-06,
1437
+ "loss": 1.498,
1438
+ "step": 204
1439
+ },
1440
+ {
1441
+ "epoch": 0.6074074074074074,
1442
+ "grad_norm": 9.75,
1443
+ "learning_rate": 7.555227423214174e-06,
1444
+ "loss": 1.4675,
1445
+ "step": 205
1446
+ },
1447
+ {
1448
+ "epoch": 0.6103703703703703,
1449
+ "grad_norm": 9.75,
1450
+ "learning_rate": 7.45955390973181e-06,
1451
+ "loss": 1.5049,
1452
+ "step": 206
1453
+ },
1454
+ {
1455
+ "epoch": 0.6133333333333333,
1456
+ "grad_norm": 9.125,
1457
+ "learning_rate": 7.364128339309326e-06,
1458
+ "loss": 1.4279,
1459
+ "step": 207
1460
+ },
1461
+ {
1462
+ "epoch": 0.6162962962962963,
1463
+ "grad_norm": 9.5625,
1464
+ "learning_rate": 7.268960025314169e-06,
1465
+ "loss": 1.504,
1466
+ "step": 208
1467
+ },
1468
+ {
1469
+ "epoch": 0.6192592592592593,
1470
+ "grad_norm": 9.0625,
1471
+ "learning_rate": 7.174058256006012e-06,
1472
+ "loss": 1.4354,
1473
+ "step": 209
1474
+ },
1475
+ {
1476
+ "epoch": 0.6222222222222222,
1477
+ "grad_norm": 9.1875,
1478
+ "learning_rate": 7.079432293630244e-06,
1479
+ "loss": 1.3362,
1480
+ "step": 210
1481
+ },
1482
+ {
1483
+ "epoch": 0.6251851851851852,
1484
+ "grad_norm": 9.9375,
1485
+ "learning_rate": 6.985091373513972e-06,
1486
+ "loss": 1.3616,
1487
+ "step": 211
1488
+ },
1489
+ {
1490
+ "epoch": 0.6281481481481481,
1491
+ "grad_norm": 9.4375,
1492
+ "learning_rate": 6.8910447031646884e-06,
1493
+ "loss": 1.4284,
1494
+ "step": 212
1495
+ },
1496
+ {
1497
+ "epoch": 0.6311111111111111,
1498
+ "grad_norm": 9.0,
1499
+ "learning_rate": 6.797301461371626e-06,
1500
+ "loss": 1.2881,
1501
+ "step": 213
1502
+ },
1503
+ {
1504
+ "epoch": 0.6340740740740741,
1505
+ "grad_norm": 10.0625,
1506
+ "learning_rate": 6.703870797309922e-06,
1507
+ "loss": 1.4838,
1508
+ "step": 214
1509
+ },
1510
+ {
1511
+ "epoch": 0.6370370370370371,
1512
+ "grad_norm": 10.625,
1513
+ "learning_rate": 6.610761829647685e-06,
1514
+ "loss": 1.5364,
1515
+ "step": 215
1516
+ },
1517
+ {
1518
+ "epoch": 0.64,
1519
+ "grad_norm": 9.625,
1520
+ "learning_rate": 6.517983645656014e-06,
1521
+ "loss": 1.4076,
1522
+ "step": 216
1523
+ },
1524
+ {
1525
+ "epoch": 0.642962962962963,
1526
+ "grad_norm": 9.8125,
1527
+ "learning_rate": 6.4255453003221115e-06,
1528
+ "loss": 1.3774,
1529
+ "step": 217
1530
+ },
1531
+ {
1532
+ "epoch": 0.6459259259259259,
1533
+ "grad_norm": 9.4375,
1534
+ "learning_rate": 6.33345581546552e-06,
1535
+ "loss": 1.4266,
1536
+ "step": 218
1537
+ },
1538
+ {
1539
+ "epoch": 0.6488888888888888,
1540
+ "grad_norm": 8.875,
1541
+ "learning_rate": 6.241724178857621e-06,
1542
+ "loss": 1.3279,
1543
+ "step": 219
1544
+ },
1545
+ {
1546
+ "epoch": 0.6518518518518519,
1547
+ "grad_norm": 10.5625,
1548
+ "learning_rate": 6.1503593433444316e-06,
1549
+ "loss": 1.4278,
1550
+ "step": 220
1551
+ },
1552
+ {
1553
+ "epoch": 0.6548148148148148,
1554
+ "grad_norm": 9.6875,
1555
+ "learning_rate": 6.059370225972834e-06,
1556
+ "loss": 1.454,
1557
+ "step": 221
1558
+ },
1559
+ {
1560
+ "epoch": 0.6577777777777778,
1561
+ "grad_norm": 9.5,
1562
+ "learning_rate": 5.96876570712028e-06,
1563
+ "loss": 1.3883,
1564
+ "step": 222
1565
+ },
1566
+ {
1567
+ "epoch": 0.6607407407407407,
1568
+ "grad_norm": 9.5,
1569
+ "learning_rate": 5.878554629628081e-06,
1570
+ "loss": 1.3739,
1571
+ "step": 223
1572
+ },
1573
+ {
1574
+ "epoch": 0.6637037037037037,
1575
+ "grad_norm": 9.3125,
1576
+ "learning_rate": 5.788745797938372e-06,
1577
+ "loss": 1.381,
1578
+ "step": 224
1579
+ },
1580
+ {
1581
+ "epoch": 0.6666666666666666,
1582
+ "grad_norm": 9.375,
1583
+ "learning_rate": 5.699347977234799e-06,
1584
+ "loss": 1.3763,
1585
+ "step": 225
1586
+ },
1587
+ {
1588
+ "epoch": 0.6696296296296296,
1589
+ "grad_norm": 9.4375,
1590
+ "learning_rate": 5.610369892587064e-06,
1591
+ "loss": 1.2782,
1592
+ "step": 226
1593
+ },
1594
+ {
1595
+ "epoch": 0.6725925925925926,
1596
+ "grad_norm": 9.1875,
1597
+ "learning_rate": 5.5218202280993725e-06,
1598
+ "loss": 1.428,
1599
+ "step": 227
1600
+ },
1601
+ {
1602
+ "epoch": 0.6755555555555556,
1603
+ "grad_norm": 9.4375,
1604
+ "learning_rate": 5.43370762606287e-06,
1605
+ "loss": 1.4381,
1606
+ "step": 228
1607
+ },
1608
+ {
1609
+ "epoch": 0.6785185185185185,
1610
+ "grad_norm": 9.875,
1611
+ "learning_rate": 5.346040686112189e-06,
1612
+ "loss": 1.3499,
1613
+ "step": 229
1614
+ },
1615
+ {
1616
+ "epoch": 0.6814814814814815,
1617
+ "grad_norm": 9.6875,
1618
+ "learning_rate": 5.25882796438612e-06,
1619
+ "loss": 1.3794,
1620
+ "step": 230
1621
+ },
1622
+ {
1623
+ "epoch": 0.6844444444444444,
1624
+ "grad_norm": 9.6875,
1625
+ "learning_rate": 5.172077972692553e-06,
1626
+ "loss": 1.3564,
1627
+ "step": 231
1628
+ },
1629
+ {
1630
+ "epoch": 0.6874074074074074,
1631
+ "grad_norm": 10.0,
1632
+ "learning_rate": 5.08579917767775e-06,
1633
+ "loss": 1.4274,
1634
+ "step": 232
1635
+ },
1636
+ {
1637
+ "epoch": 0.6903703703703704,
1638
+ "grad_norm": 8.9375,
1639
+ "learning_rate": 5.000000000000003e-06,
1640
+ "loss": 1.267,
1641
+ "step": 233
1642
+ },
1643
+ {
1644
+ "epoch": 0.6933333333333334,
1645
+ "grad_norm": 9.375,
1646
+ "learning_rate": 4.914688813507798e-06,
1647
+ "loss": 1.3242,
1648
+ "step": 234
1649
+ },
1650
+ {
1651
+ "epoch": 0.6962962962962963,
1652
+ "grad_norm": 9.25,
1653
+ "learning_rate": 4.829873944422544e-06,
1654
+ "loss": 1.3488,
1655
+ "step": 235
1656
+ },
1657
+ {
1658
+ "epoch": 0.6992592592592592,
1659
+ "grad_norm": 9.125,
1660
+ "learning_rate": 4.745563670525942e-06,
1661
+ "loss": 1.3237,
1662
+ "step": 236
1663
+ },
1664
+ {
1665
+ "epoch": 0.7022222222222222,
1666
+ "grad_norm": 9.5625,
1667
+ "learning_rate": 4.661766220352098e-06,
1668
+ "loss": 1.3183,
1669
+ "step": 237
1670
+ },
1671
+ {
1672
+ "epoch": 0.7051851851851851,
1673
+ "grad_norm": 9.5625,
1674
+ "learning_rate": 4.578489772384415e-06,
1675
+ "loss": 1.4587,
1676
+ "step": 238
1677
+ },
1678
+ {
1679
+ "epoch": 0.7081481481481482,
1680
+ "grad_norm": 9.5,
1681
+ "learning_rate": 4.495742454257418e-06,
1682
+ "loss": 1.2974,
1683
+ "step": 239
1684
+ },
1685
+ {
1686
+ "epoch": 0.7111111111111111,
1687
+ "grad_norm": 9.5625,
1688
+ "learning_rate": 4.413532341963477e-06,
1689
+ "loss": 1.4561,
1690
+ "step": 240
1691
+ },
1692
+ {
1693
+ "epoch": 0.7140740740740741,
1694
+ "grad_norm": 9.625,
1695
+ "learning_rate": 4.331867459064623e-06,
1696
+ "loss": 1.4047,
1697
+ "step": 241
1698
+ },
1699
+ {
1700
+ "epoch": 0.717037037037037,
1701
+ "grad_norm": 9.3125,
1702
+ "learning_rate": 4.250755775909465e-06,
1703
+ "loss": 1.4223,
1704
+ "step": 242
1705
+ },
1706
+ {
1707
+ "epoch": 0.72,
1708
+ "grad_norm": 9.3125,
1709
+ "learning_rate": 4.170205208855281e-06,
1710
+ "loss": 1.4031,
1711
+ "step": 243
1712
+ },
1713
+ {
1714
+ "epoch": 0.7229629629629629,
1715
+ "grad_norm": 9.125,
1716
+ "learning_rate": 4.090223619495419e-06,
1717
+ "loss": 1.3283,
1718
+ "step": 244
1719
+ },
1720
+ {
1721
+ "epoch": 0.725925925925926,
1722
+ "grad_norm": 9.375,
1723
+ "learning_rate": 4.010818813892e-06,
1724
+ "loss": 1.419,
1725
+ "step": 245
1726
+ },
1727
+ {
1728
+ "epoch": 0.7288888888888889,
1729
+ "grad_norm": 9.25,
1730
+ "learning_rate": 3.931998541814069e-06,
1731
+ "loss": 1.3429,
1732
+ "step": 246
1733
+ },
1734
+ {
1735
+ "epoch": 0.7318518518518519,
1736
+ "grad_norm": 9.3125,
1737
+ "learning_rate": 3.85377049598123e-06,
1738
+ "loss": 1.436,
1739
+ "step": 247
1740
+ },
1741
+ {
1742
+ "epoch": 0.7348148148148148,
1743
+ "grad_norm": 9.375,
1744
+ "learning_rate": 3.7761423113128427e-06,
1745
+ "loss": 1.3918,
1746
+ "step": 248
1747
+ },
1748
+ {
1749
+ "epoch": 0.7377777777777778,
1750
+ "grad_norm": 9.6875,
1751
+ "learning_rate": 3.6991215641828903e-06,
1752
+ "loss": 1.3523,
1753
+ "step": 249
1754
+ },
1755
+ {
1756
+ "epoch": 0.7407407407407407,
1757
+ "grad_norm": 9.4375,
1758
+ "learning_rate": 3.622715771680508e-06,
1759
+ "loss": 1.3808,
1760
+ "step": 250
1761
+ },
1762
+ {
1763
+ "epoch": 0.7437037037037038,
1764
+ "grad_norm": 9.375,
1765
+ "learning_rate": 3.5469323908763507e-06,
1766
+ "loss": 1.4308,
1767
+ "step": 251
1768
+ },
1769
+ {
1770
+ "epoch": 0.7466666666666667,
1771
+ "grad_norm": 9.4375,
1772
+ "learning_rate": 3.4717788180947855e-06,
1773
+ "loss": 1.3985,
1774
+ "step": 252
1775
+ },
1776
+ {
1777
+ "epoch": 0.7496296296296296,
1778
+ "grad_norm": 9.625,
1779
+ "learning_rate": 3.3972623881920296e-06,
1780
+ "loss": 1.3034,
1781
+ "step": 253
1782
+ },
1783
+ {
1784
+ "epoch": 0.7525925925925926,
1785
+ "grad_norm": 9.3125,
1786
+ "learning_rate": 3.323390373840276e-06,
1787
+ "loss": 1.3579,
1788
+ "step": 254
1789
+ },
1790
+ {
1791
+ "epoch": 0.7555555555555555,
1792
+ "grad_norm": 9.25,
1793
+ "learning_rate": 3.250169984817897e-06,
1794
+ "loss": 1.3301,
1795
+ "step": 255
1796
+ },
1797
+ {
1798
+ "epoch": 0.7585185185185185,
1799
+ "grad_norm": 9.125,
1800
+ "learning_rate": 3.1776083673057834e-06,
1801
+ "loss": 1.308,
1802
+ "step": 256
1803
+ },
1804
+ {
1805
+ "epoch": 0.7614814814814815,
1806
+ "grad_norm": 9.1875,
1807
+ "learning_rate": 3.1057126031898843e-06,
1808
+ "loss": 1.3278,
1809
+ "step": 257
1810
+ },
1811
+ {
1812
+ "epoch": 0.7644444444444445,
1813
+ "grad_norm": 9.4375,
1814
+ "learning_rate": 3.0344897093700333e-06,
1815
+ "loss": 1.3678,
1816
+ "step": 258
1817
+ },
1818
+ {
1819
+ "epoch": 0.7674074074074074,
1820
+ "grad_norm": 9.25,
1821
+ "learning_rate": 2.963946637075107e-06,
1822
+ "loss": 1.2888,
1823
+ "step": 259
1824
+ },
1825
+ {
1826
+ "epoch": 0.7703703703703704,
1827
+ "grad_norm": 9.1875,
1828
+ "learning_rate": 2.8940902711846052e-06,
1829
+ "loss": 1.2899,
1830
+ "step": 260
1831
+ },
1832
+ {
1833
+ "epoch": 0.7733333333333333,
1834
+ "grad_norm": 8.625,
1835
+ "learning_rate": 2.8249274295566863e-06,
1836
+ "loss": 1.2525,
1837
+ "step": 261
1838
+ },
1839
+ {
1840
+ "epoch": 0.7762962962962963,
1841
+ "grad_norm": 8.8125,
1842
+ "learning_rate": 2.756464862362772e-06,
1843
+ "loss": 1.2665,
1844
+ "step": 262
1845
+ },
1846
+ {
1847
+ "epoch": 0.7792592592592592,
1848
+ "grad_norm": 9.4375,
1849
+ "learning_rate": 2.688709251428725e-06,
1850
+ "loss": 1.363,
1851
+ "step": 263
1852
+ },
1853
+ {
1854
+ "epoch": 0.7822222222222223,
1855
+ "grad_norm": 9.0625,
1856
+ "learning_rate": 2.6216672095827267e-06,
1857
+ "loss": 1.3086,
1858
+ "step": 264
1859
+ },
1860
+ {
1861
+ "epoch": 0.7851851851851852,
1862
+ "grad_norm": 9.3125,
1863
+ "learning_rate": 2.555345280009872e-06,
1864
+ "loss": 1.4259,
1865
+ "step": 265
1866
+ },
1867
+ {
1868
+ "epoch": 0.7881481481481482,
1869
+ "grad_norm": 9.125,
1870
+ "learning_rate": 2.4897499356135712e-06,
1871
+ "loss": 1.2748,
1872
+ "step": 266
1873
+ },
1874
+ {
1875
+ "epoch": 0.7911111111111111,
1876
+ "grad_norm": 9.25,
1877
+ "learning_rate": 2.424887578383799e-06,
1878
+ "loss": 1.329,
1879
+ "step": 267
1880
+ },
1881
+ {
1882
+ "epoch": 0.794074074074074,
1883
+ "grad_norm": 9.5625,
1884
+ "learning_rate": 2.3607645387722753e-06,
1885
+ "loss": 1.4323,
1886
+ "step": 268
1887
+ },
1888
+ {
1889
+ "epoch": 0.797037037037037,
1890
+ "grad_norm": 9.5625,
1891
+ "learning_rate": 2.2973870750746253e-06,
1892
+ "loss": 1.3081,
1893
+ "step": 269
1894
+ },
1895
+ {
1896
+ "epoch": 0.8,
1897
+ "grad_norm": 9.375,
1898
+ "learning_rate": 2.234761372819577e-06,
1899
+ "loss": 1.3967,
1900
+ "step": 270
1901
+ },
1902
+ {
1903
+ "epoch": 0.802962962962963,
1904
+ "grad_norm": 9.3125,
1905
+ "learning_rate": 2.1728935441652687e-06,
1906
+ "loss": 1.345,
1907
+ "step": 271
1908
+ },
1909
+ {
1910
+ "epoch": 0.8059259259259259,
1911
+ "grad_norm": 9.0,
1912
+ "learning_rate": 2.111789627302707e-06,
1913
+ "loss": 1.3009,
1914
+ "step": 272
1915
+ },
1916
+ {
1917
+ "epoch": 0.8088888888888889,
1918
+ "grad_norm": 9.25,
1919
+ "learning_rate": 2.0514555858664663e-06,
1920
+ "loss": 1.3242,
1921
+ "step": 273
1922
+ },
1923
+ {
1924
+ "epoch": 0.8118518518518518,
1925
+ "grad_norm": 9.3125,
1926
+ "learning_rate": 1.991897308352624e-06,
1927
+ "loss": 1.3109,
1928
+ "step": 274
1929
+ },
1930
+ {
1931
+ "epoch": 0.8148148148148148,
1932
+ "grad_norm": 8.9375,
1933
+ "learning_rate": 1.93312060754407e-06,
1934
+ "loss": 1.3353,
1935
+ "step": 275
1936
+ },
1937
+ {
1938
+ "epoch": 0.8177777777777778,
1939
+ "grad_norm": 9.625,
1940
+ "learning_rate": 1.875131219943187e-06,
1941
+ "loss": 1.2318,
1942
+ "step": 276
1943
+ },
1944
+ {
1945
+ "epoch": 0.8207407407407408,
1946
+ "grad_norm": 9.0,
1947
+ "learning_rate": 1.817934805211976e-06,
1948
+ "loss": 1.2841,
1949
+ "step": 277
1950
+ },
1951
+ {
1952
+ "epoch": 0.8237037037037037,
1953
+ "grad_norm": 8.6875,
1954
+ "learning_rate": 1.761536945619684e-06,
1955
+ "loss": 1.2261,
1956
+ "step": 278
1957
+ },
1958
+ {
1959
+ "epoch": 0.8266666666666667,
1960
+ "grad_norm": 9.5625,
1961
+ "learning_rate": 1.7059431454979825e-06,
1962
+ "loss": 1.364,
1963
+ "step": 279
1964
+ },
1965
+ {
1966
+ "epoch": 0.8296296296296296,
1967
+ "grad_norm": 9.125,
1968
+ "learning_rate": 1.6511588307037596e-06,
1969
+ "loss": 1.426,
1970
+ "step": 280
1971
+ },
1972
+ {
1973
+ "epoch": 0.8325925925925926,
1974
+ "grad_norm": 9.625,
1975
+ "learning_rate": 1.5971893480895583e-06,
1976
+ "loss": 1.458,
1977
+ "step": 281
1978
+ },
1979
+ {
1980
+ "epoch": 0.8355555555555556,
1981
+ "grad_norm": 8.875,
1982
+ "learning_rate": 1.5440399649817384e-06,
1983
+ "loss": 1.4178,
1984
+ "step": 282
1985
+ },
1986
+ {
1987
+ "epoch": 0.8385185185185186,
1988
+ "grad_norm": 9.0625,
1989
+ "learning_rate": 1.4917158686663992e-06,
1990
+ "loss": 1.3404,
1991
+ "step": 283
1992
+ },
1993
+ {
1994
+ "epoch": 0.8414814814814815,
1995
+ "grad_norm": 8.75,
1996
+ "learning_rate": 1.4402221658830963e-06,
1997
+ "loss": 1.3145,
1998
+ "step": 284
1999
+ },
2000
+ {
2001
+ "epoch": 0.8444444444444444,
2002
+ "grad_norm": 9.3125,
2003
+ "learning_rate": 1.3895638823264447e-06,
2004
+ "loss": 1.2529,
2005
+ "step": 285
2006
+ },
2007
+ {
2008
+ "epoch": 0.8474074074074074,
2009
+ "grad_norm": 10.6875,
2010
+ "learning_rate": 1.339745962155613e-06,
2011
+ "loss": 1.5298,
2012
+ "step": 286
2013
+ },
2014
+ {
2015
+ "epoch": 0.8503703703703703,
2016
+ "grad_norm": 8.75,
2017
+ "learning_rate": 1.2907732675117878e-06,
2018
+ "loss": 1.2976,
2019
+ "step": 287
2020
+ },
2021
+ {
2022
+ "epoch": 0.8533333333333334,
2023
+ "grad_norm": 9.875,
2024
+ "learning_rate": 1.2426505780436326e-06,
2025
+ "loss": 1.3695,
2026
+ "step": 288
2027
+ },
2028
+ {
2029
+ "epoch": 0.8562962962962963,
2030
+ "grad_norm": 8.75,
2031
+ "learning_rate": 1.1953825904408033e-06,
2032
+ "loss": 1.2496,
2033
+ "step": 289
2034
+ },
2035
+ {
2036
+ "epoch": 0.8592592592592593,
2037
+ "grad_norm": 9.8125,
2038
+ "learning_rate": 1.1489739179755622e-06,
2039
+ "loss": 1.4167,
2040
+ "step": 290
2041
+ },
2042
+ {
2043
+ "epoch": 0.8622222222222222,
2044
+ "grad_norm": 9.5,
2045
+ "learning_rate": 1.1034290900525279e-06,
2046
+ "loss": 1.3485,
2047
+ "step": 291
2048
+ },
2049
+ {
2050
+ "epoch": 0.8651851851851852,
2051
+ "grad_norm": 8.9375,
2052
+ "learning_rate": 1.0587525517666142e-06,
2053
+ "loss": 1.2987,
2054
+ "step": 292
2055
+ },
2056
+ {
2057
+ "epoch": 0.8681481481481481,
2058
+ "grad_norm": 9.3125,
2059
+ "learning_rate": 1.0149486634692019e-06,
2060
+ "loss": 1.2708,
2061
+ "step": 293
2062
+ },
2063
+ {
2064
+ "epoch": 0.8711111111111111,
2065
+ "grad_norm": 9.0625,
2066
+ "learning_rate": 9.720217003425648e-07,
2067
+ "loss": 1.405,
2068
+ "step": 294
2069
+ },
2070
+ {
2071
+ "epoch": 0.8740740740740741,
2072
+ "grad_norm": 9.125,
2073
+ "learning_rate": 9.299758519826274e-07,
2074
+ "loss": 1.3643,
2075
+ "step": 295
2076
+ },
2077
+ {
2078
+ "epoch": 0.8770370370370371,
2079
+ "grad_norm": 9.125,
2080
+ "learning_rate": 8.888152219900692e-07,
2081
+ "loss": 1.33,
2082
+ "step": 296
2083
+ },
2084
+ {
2085
+ "epoch": 0.88,
2086
+ "grad_norm": 9.6875,
2087
+ "learning_rate": 8.485438275698154e-07,
2088
+ "loss": 1.4734,
2089
+ "step": 297
2090
+ },
2091
+ {
2092
+ "epoch": 0.882962962962963,
2093
+ "grad_norm": 9.75,
2094
+ "learning_rate": 8.091655991389668e-07,
2095
+ "loss": 1.3231,
2096
+ "step": 298
2097
+ },
2098
+ {
2099
+ "epoch": 0.8859259259259259,
2100
+ "grad_norm": 8.9375,
2101
+ "learning_rate": 7.706843799431985e-07,
2102
+ "loss": 1.2256,
2103
+ "step": 299
2104
+ },
2105
+ {
2106
+ "epoch": 0.8888888888888888,
2107
+ "grad_norm": 9.6875,
2108
+ "learning_rate": 7.331039256816664e-07,
2109
+ "loss": 1.3831,
2110
+ "step": 300
2111
+ },
2112
+ {
2113
+ "epoch": 0.8918518518518519,
2114
+ "grad_norm": 8.9375,
2115
+ "learning_rate": 6.964279041404553e-07,
2116
+ "loss": 1.2743,
2117
+ "step": 301
2118
+ },
2119
+ {
2120
+ "epoch": 0.8948148148148148,
2121
+ "grad_norm": 9.1875,
2122
+ "learning_rate": 6.606598948346132e-07,
2123
+ "loss": 1.3141,
2124
+ "step": 302
2125
+ },
2126
+ {
2127
+ "epoch": 0.8977777777777778,
2128
+ "grad_norm": 9.8125,
2129
+ "learning_rate": 6.258033886587911e-07,
2130
+ "loss": 1.4539,
2131
+ "step": 303
2132
+ },
2133
+ {
2134
+ "epoch": 0.9007407407407407,
2135
+ "grad_norm": 10.3125,
2136
+ "learning_rate": 5.918617875465449e-07,
2137
+ "loss": 1.4409,
2138
+ "step": 304
2139
+ },
2140
+ {
2141
+ "epoch": 0.9037037037037037,
2142
+ "grad_norm": 9.9375,
2143
+ "learning_rate": 5.588384041383089e-07,
2144
+ "loss": 1.3793,
2145
+ "step": 305
2146
+ },
2147
+ {
2148
+ "epoch": 0.9066666666666666,
2149
+ "grad_norm": 9.8125,
2150
+ "learning_rate": 5.267364614580861e-07,
2151
+ "loss": 1.391,
2152
+ "step": 306
2153
+ },
2154
+ {
2155
+ "epoch": 0.9096296296296297,
2156
+ "grad_norm": 8.9375,
2157
+ "learning_rate": 4.955590925988896e-07,
2158
+ "loss": 1.2844,
2159
+ "step": 307
2160
+ },
2161
+ {
2162
+ "epoch": 0.9125925925925926,
2163
+ "grad_norm": 9.6875,
2164
+ "learning_rate": 4.653093404169573e-07,
2165
+ "loss": 1.3854,
2166
+ "step": 308
2167
+ },
2168
+ {
2169
+ "epoch": 0.9155555555555556,
2170
+ "grad_norm": 9.9375,
2171
+ "learning_rate": 4.359901572347758e-07,
2172
+ "loss": 1.378,
2173
+ "step": 309
2174
+ },
2175
+ {
2176
+ "epoch": 0.9185185185185185,
2177
+ "grad_norm": 10.0625,
2178
+ "learning_rate": 4.0760440455293703e-07,
2179
+ "loss": 1.4957,
2180
+ "step": 310
2181
+ },
2182
+ {
2183
+ "epoch": 0.9214814814814815,
2184
+ "grad_norm": 10.0625,
2185
+ "learning_rate": 3.801548527708621e-07,
2186
+ "loss": 1.518,
2187
+ "step": 311
2188
+ },
2189
+ {
2190
+ "epoch": 0.9244444444444444,
2191
+ "grad_norm": 9.1875,
2192
+ "learning_rate": 3.5364418091641374e-07,
2193
+ "loss": 1.2608,
2194
+ "step": 312
2195
+ },
2196
+ {
2197
+ "epoch": 0.9274074074074075,
2198
+ "grad_norm": 9.5625,
2199
+ "learning_rate": 3.280749763844293e-07,
2200
+ "loss": 1.3896,
2201
+ "step": 313
2202
+ },
2203
+ {
2204
+ "epoch": 0.9303703703703704,
2205
+ "grad_norm": 9.4375,
2206
+ "learning_rate": 3.034497346841958e-07,
2207
+ "loss": 1.4369,
2208
+ "step": 314
2209
+ },
2210
+ {
2211
+ "epoch": 0.9333333333333333,
2212
+ "grad_norm": 9.0,
2213
+ "learning_rate": 2.7977085919589253e-07,
2214
+ "loss": 1.4707,
2215
+ "step": 315
2216
+ },
2217
+ {
2218
+ "epoch": 0.9362962962962963,
2219
+ "grad_norm": 8.875,
2220
+ "learning_rate": 2.570406609360221e-07,
2221
+ "loss": 1.2772,
2222
+ "step": 316
2223
+ },
2224
+ {
2225
+ "epoch": 0.9392592592592592,
2226
+ "grad_norm": 9.6875,
2227
+ "learning_rate": 2.3526135833186527e-07,
2228
+ "loss": 1.3784,
2229
+ "step": 317
2230
+ },
2231
+ {
2232
+ "epoch": 0.9422222222222222,
2233
+ "grad_norm": 9.3125,
2234
+ "learning_rate": 2.1443507700495968e-07,
2235
+ "loss": 1.3823,
2236
+ "step": 318
2237
+ },
2238
+ {
2239
+ "epoch": 0.9451851851851852,
2240
+ "grad_norm": 9.375,
2241
+ "learning_rate": 1.9456384956365149e-07,
2242
+ "loss": 1.3478,
2243
+ "step": 319
2244
+ },
2245
+ {
2246
+ "epoch": 0.9481481481481482,
2247
+ "grad_norm": 9.875,
2248
+ "learning_rate": 1.7564961540470492e-07,
2249
+ "loss": 1.3708,
2250
+ "step": 320
2251
+ },
2252
+ {
2253
+ "epoch": 0.9511111111111111,
2254
+ "grad_norm": 9.125,
2255
+ "learning_rate": 1.5769422052403172e-07,
2256
+ "loss": 1.2729,
2257
+ "step": 321
2258
+ },
2259
+ {
2260
+ "epoch": 0.9540740740740741,
2261
+ "grad_norm": 9.3125,
2262
+ "learning_rate": 1.4069941733651948e-07,
2263
+ "loss": 1.3488,
2264
+ "step": 322
2265
+ },
2266
+ {
2267
+ "epoch": 0.957037037037037,
2268
+ "grad_norm": 9.0,
2269
+ "learning_rate": 1.2466686450499866e-07,
2270
+ "loss": 1.3308,
2271
+ "step": 323
2272
+ },
2273
+ {
2274
+ "epoch": 0.96,
2275
+ "grad_norm": 9.1875,
2276
+ "learning_rate": 1.0959812677835968e-07,
2277
+ "loss": 1.3702,
2278
+ "step": 324
2279
+ },
2280
+ {
2281
+ "epoch": 0.9629629629629629,
2282
+ "grad_norm": 9.125,
2283
+ "learning_rate": 9.549467483884412e-08,
2284
+ "loss": 1.2894,
2285
+ "step": 325
2286
+ },
2287
+ {
2288
+ "epoch": 0.965925925925926,
2289
+ "grad_norm": 9.6875,
2290
+ "learning_rate": 8.235788515849607e-08,
2291
+ "loss": 1.4597,
2292
+ "step": 326
2293
+ },
2294
+ {
2295
+ "epoch": 0.9688888888888889,
2296
+ "grad_norm": 8.9375,
2297
+ "learning_rate": 7.018903986483083e-08,
2298
+ "loss": 1.3452,
2299
+ "step": 327
2300
+ },
2301
+ {
2302
+ "epoch": 0.9718518518518519,
2303
+ "grad_norm": 9.1875,
2304
+ "learning_rate": 5.89893266156949e-08,
2305
+ "loss": 1.3317,
2306
+ "step": 328
2307
+ },
2308
+ {
2309
+ "epoch": 0.9748148148148148,
2310
+ "grad_norm": 9.9375,
2311
+ "learning_rate": 4.8759838483358745e-08,
2312
+ "loss": 1.325,
2313
+ "step": 329
2314
+ },
2315
+ {
2316
+ "epoch": 0.9777777777777777,
2317
+ "grad_norm": 9.125,
2318
+ "learning_rate": 3.950157384783104e-08,
2319
+ "loss": 1.3164,
2320
+ "step": 330
2321
+ },
2322
+ {
2323
+ "epoch": 0.9807407407407407,
2324
+ "grad_norm": 9.4375,
2325
+ "learning_rate": 3.12154362994177e-08,
2326
+ "loss": 1.4195,
2327
+ "step": 331
2328
+ },
2329
+ {
2330
+ "epoch": 0.9837037037037037,
2331
+ "grad_norm": 9.375,
2332
+ "learning_rate": 2.3902234550536862e-08,
2333
+ "loss": 1.3961,
2334
+ "step": 332
2335
+ },
2336
+ {
2337
+ "epoch": 0.9866666666666667,
2338
+ "grad_norm": 9.1875,
2339
+ "learning_rate": 1.7562682356786488e-08,
2340
+ "loss": 1.2901,
2341
+ "step": 333
2342
+ },
2343
+ {
2344
+ "epoch": 0.9896296296296296,
2345
+ "grad_norm": 10.0,
2346
+ "learning_rate": 1.2197398447283404e-08,
2347
+ "loss": 1.5034,
2348
+ "step": 334
2349
+ },
2350
+ {
2351
+ "epoch": 0.9925925925925926,
2352
+ "grad_norm": 10.0625,
2353
+ "learning_rate": 7.806906464281617e-09,
2354
+ "loss": 1.4242,
2355
+ "step": 335
2356
+ },
2357
+ {
2358
+ "epoch": 0.9955555555555555,
2359
+ "grad_norm": 9.625,
2360
+ "learning_rate": 4.39163491205652e-09,
2361
+ "loss": 1.4248,
2362
+ "step": 336
2363
+ },
2364
+ {
2365
+ "epoch": 0.9985185185185185,
2366
+ "grad_norm": 9.8125,
2367
+ "learning_rate": 1.951917115091684e-09,
2368
+ "loss": 1.3997,
2369
+ "step": 337
2370
+ },
2371
+ {
2372
+ "epoch": 1.0,
2373
+ "grad_norm": 13.5625,
2374
+ "learning_rate": 4.879911855426578e-10,
2375
+ "loss": 1.3577,
2376
+ "step": 338
2377
+ }
2378
+ ],
2379
+ "logging_steps": 1,
2380
+ "max_steps": 338,
2381
+ "num_input_tokens_seen": 0,
2382
+ "num_train_epochs": 1,
2383
+ "save_steps": 10000,
2384
+ "stateful_callbacks": {
2385
+ "TrainerControl": {
2386
+ "args": {
2387
+ "should_epoch_stop": false,
2388
+ "should_evaluate": false,
2389
+ "should_log": false,
2390
+ "should_save": true,
2391
+ "should_training_stop": true
2392
+ },
2393
+ "attributes": {}
2394
+ }
2395
+ },
2396
+ "total_flos": 5614551709925376.0,
2397
+ "train_batch_size": 2,
2398
+ "trial_name": null,
2399
+ "trial_params": null
2400
+ }
checkpoint-338/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9e208ca8d5ec9994728b931eda9cdb80dc0a6ebdb8d0187175a9d1fd08f1ba1
3
+ size 5752