Ekanari commited on
Commit
6ad514d
·
verified ·
1 Parent(s): 6dcef89

Upload folder using huggingface_hub

Browse files
sft_output_simpleQA/VNLsimpleQA_test_predictions.json ADDED
The diff for this file is too large to render. See raw diff
 
sft_output_simpleQA/checkpoint-2508/config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MptForCausalLM"
4
+ ],
5
+ "attn_config": {
6
+ "alibi": true,
7
+ "alibi_bias_max": 8,
8
+ "attn_impl": "torch",
9
+ "attn_pdrop": 0.0,
10
+ "attn_type": "multihead_attention",
11
+ "attn_uses_sequence_id": false,
12
+ "clip_qkv": null,
13
+ "model_type": "",
14
+ "prefix_lm": false,
15
+ "qk_ln": false,
16
+ "softmax_scale": null,
17
+ "torch_dtype": "bfloat16"
18
+ },
19
+ "auto_map": {
20
+ "AutoConfig": "vinai/PhoGPT-4B-Chat--configuration_mpt.MPTConfig",
21
+ "AutoModelForCausalLM": "vinai/PhoGPT-4B-Chat--modeling_mpt.MPTForCausalLM"
22
+ },
23
+ "d_model": 3072,
24
+ "emb_pdrop": 0.0,
25
+ "embedding_fraction": 1.0,
26
+ "expansion_ratio": 4,
27
+ "init_config": {
28
+ "emb_init_std": null,
29
+ "emb_init_uniform_lim": null,
30
+ "fan_mode": "fan_in",
31
+ "init_div_is_residual": true,
32
+ "init_gain": 0.0,
33
+ "init_nonlinearity": "relu",
34
+ "init_std": null,
35
+ "name": "kaiming_normal_",
36
+ "verbose": 0
37
+ },
38
+ "init_device": "cpu",
39
+ "initializer_range": 0.02,
40
+ "layer_norm_epsilon": 1e-05,
41
+ "learned_pos_emb": true,
42
+ "logit_scale": null,
43
+ "max_seq_len": 8192,
44
+ "model_type": "mpt",
45
+ "n_heads": 24,
46
+ "n_layers": 32,
47
+ "no_bias": false,
48
+ "norm_type": "low_precision_layernorm",
49
+ "resid_pdrop": 0.0,
50
+ "torch_dtype": "bfloat16",
51
+ "transformers_version": "4.51.3",
52
+ "use_cache": false,
53
+ "verbose": 0,
54
+ "vocab_size": 20480
55
+ }
sft_output_simpleQA/checkpoint-2508/generation_config.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.51.3",
4
+ "use_cache": false
5
+ }
sft_output_simpleQA/checkpoint-2508/model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:914f85135733d4c80639657e3889ebaa438af4bd2ddda22e0b4afde39183028a
3
+ size 4957952800
sft_output_simpleQA/checkpoint-2508/model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aedaba4b108fa3668dfcc6736f0b71c640c3acab1674ef9ab79e84ed15b7c55b
3
+ size 2416055432
sft_output_simpleQA/checkpoint-2508/model.safetensors.index.json ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 7373985792
4
+ },
5
+ "weight_map": {
6
+ "transformer.blocks.0.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
7
+ "transformer.blocks.0.attn.out_proj.weight": "model-00001-of-00002.safetensors",
8
+ "transformer.blocks.0.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
9
+ "transformer.blocks.0.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
10
+ "transformer.blocks.0.norm_1.weight": "model-00001-of-00002.safetensors",
11
+ "transformer.blocks.0.norm_2.weight": "model-00001-of-00002.safetensors",
12
+ "transformer.blocks.1.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
13
+ "transformer.blocks.1.attn.out_proj.weight": "model-00001-of-00002.safetensors",
14
+ "transformer.blocks.1.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
15
+ "transformer.blocks.1.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
16
+ "transformer.blocks.1.norm_1.weight": "model-00001-of-00002.safetensors",
17
+ "transformer.blocks.1.norm_2.weight": "model-00001-of-00002.safetensors",
18
+ "transformer.blocks.10.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
19
+ "transformer.blocks.10.attn.out_proj.weight": "model-00001-of-00002.safetensors",
20
+ "transformer.blocks.10.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
21
+ "transformer.blocks.10.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
22
+ "transformer.blocks.10.norm_1.weight": "model-00001-of-00002.safetensors",
23
+ "transformer.blocks.10.norm_2.weight": "model-00001-of-00002.safetensors",
24
+ "transformer.blocks.11.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
25
+ "transformer.blocks.11.attn.out_proj.weight": "model-00001-of-00002.safetensors",
26
+ "transformer.blocks.11.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
27
+ "transformer.blocks.11.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
28
+ "transformer.blocks.11.norm_1.weight": "model-00001-of-00002.safetensors",
29
+ "transformer.blocks.11.norm_2.weight": "model-00001-of-00002.safetensors",
30
+ "transformer.blocks.12.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
31
+ "transformer.blocks.12.attn.out_proj.weight": "model-00001-of-00002.safetensors",
32
+ "transformer.blocks.12.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
33
+ "transformer.blocks.12.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
34
+ "transformer.blocks.12.norm_1.weight": "model-00001-of-00002.safetensors",
35
+ "transformer.blocks.12.norm_2.weight": "model-00001-of-00002.safetensors",
36
+ "transformer.blocks.13.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
37
+ "transformer.blocks.13.attn.out_proj.weight": "model-00001-of-00002.safetensors",
38
+ "transformer.blocks.13.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
39
+ "transformer.blocks.13.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
40
+ "transformer.blocks.13.norm_1.weight": "model-00001-of-00002.safetensors",
41
+ "transformer.blocks.13.norm_2.weight": "model-00001-of-00002.safetensors",
42
+ "transformer.blocks.14.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
43
+ "transformer.blocks.14.attn.out_proj.weight": "model-00001-of-00002.safetensors",
44
+ "transformer.blocks.14.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
45
+ "transformer.blocks.14.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
46
+ "transformer.blocks.14.norm_1.weight": "model-00001-of-00002.safetensors",
47
+ "transformer.blocks.14.norm_2.weight": "model-00001-of-00002.safetensors",
48
+ "transformer.blocks.15.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
49
+ "transformer.blocks.15.attn.out_proj.weight": "model-00001-of-00002.safetensors",
50
+ "transformer.blocks.15.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
51
+ "transformer.blocks.15.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
52
+ "transformer.blocks.15.norm_1.weight": "model-00001-of-00002.safetensors",
53
+ "transformer.blocks.15.norm_2.weight": "model-00001-of-00002.safetensors",
54
+ "transformer.blocks.16.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
55
+ "transformer.blocks.16.attn.out_proj.weight": "model-00001-of-00002.safetensors",
56
+ "transformer.blocks.16.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
57
+ "transformer.blocks.16.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
58
+ "transformer.blocks.16.norm_1.weight": "model-00001-of-00002.safetensors",
59
+ "transformer.blocks.16.norm_2.weight": "model-00001-of-00002.safetensors",
60
+ "transformer.blocks.17.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
61
+ "transformer.blocks.17.attn.out_proj.weight": "model-00001-of-00002.safetensors",
62
+ "transformer.blocks.17.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
63
+ "transformer.blocks.17.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
64
+ "transformer.blocks.17.norm_1.weight": "model-00001-of-00002.safetensors",
65
+ "transformer.blocks.17.norm_2.weight": "model-00001-of-00002.safetensors",
66
+ "transformer.blocks.18.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
67
+ "transformer.blocks.18.attn.out_proj.weight": "model-00001-of-00002.safetensors",
68
+ "transformer.blocks.18.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
69
+ "transformer.blocks.18.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
70
+ "transformer.blocks.18.norm_1.weight": "model-00001-of-00002.safetensors",
71
+ "transformer.blocks.18.norm_2.weight": "model-00001-of-00002.safetensors",
72
+ "transformer.blocks.19.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
73
+ "transformer.blocks.19.attn.out_proj.weight": "model-00001-of-00002.safetensors",
74
+ "transformer.blocks.19.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
75
+ "transformer.blocks.19.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
76
+ "transformer.blocks.19.norm_1.weight": "model-00001-of-00002.safetensors",
77
+ "transformer.blocks.19.norm_2.weight": "model-00001-of-00002.safetensors",
78
+ "transformer.blocks.2.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
79
+ "transformer.blocks.2.attn.out_proj.weight": "model-00001-of-00002.safetensors",
80
+ "transformer.blocks.2.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
81
+ "transformer.blocks.2.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
82
+ "transformer.blocks.2.norm_1.weight": "model-00001-of-00002.safetensors",
83
+ "transformer.blocks.2.norm_2.weight": "model-00001-of-00002.safetensors",
84
+ "transformer.blocks.20.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
85
+ "transformer.blocks.20.attn.out_proj.weight": "model-00001-of-00002.safetensors",
86
+ "transformer.blocks.20.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
87
+ "transformer.blocks.20.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
88
+ "transformer.blocks.20.norm_1.weight": "model-00001-of-00002.safetensors",
89
+ "transformer.blocks.20.norm_2.weight": "model-00001-of-00002.safetensors",
90
+ "transformer.blocks.21.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
91
+ "transformer.blocks.21.attn.out_proj.weight": "model-00001-of-00002.safetensors",
92
+ "transformer.blocks.21.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
93
+ "transformer.blocks.21.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
94
+ "transformer.blocks.21.norm_1.weight": "model-00001-of-00002.safetensors",
95
+ "transformer.blocks.21.norm_2.weight": "model-00001-of-00002.safetensors",
96
+ "transformer.blocks.22.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
97
+ "transformer.blocks.22.attn.out_proj.weight": "model-00002-of-00002.safetensors",
98
+ "transformer.blocks.22.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
99
+ "transformer.blocks.22.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
100
+ "transformer.blocks.22.norm_1.weight": "model-00002-of-00002.safetensors",
101
+ "transformer.blocks.22.norm_2.weight": "model-00002-of-00002.safetensors",
102
+ "transformer.blocks.23.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
103
+ "transformer.blocks.23.attn.out_proj.weight": "model-00002-of-00002.safetensors",
104
+ "transformer.blocks.23.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
105
+ "transformer.blocks.23.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
106
+ "transformer.blocks.23.norm_1.weight": "model-00002-of-00002.safetensors",
107
+ "transformer.blocks.23.norm_2.weight": "model-00002-of-00002.safetensors",
108
+ "transformer.blocks.24.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
109
+ "transformer.blocks.24.attn.out_proj.weight": "model-00002-of-00002.safetensors",
110
+ "transformer.blocks.24.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
111
+ "transformer.blocks.24.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
112
+ "transformer.blocks.24.norm_1.weight": "model-00002-of-00002.safetensors",
113
+ "transformer.blocks.24.norm_2.weight": "model-00002-of-00002.safetensors",
114
+ "transformer.blocks.25.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
115
+ "transformer.blocks.25.attn.out_proj.weight": "model-00002-of-00002.safetensors",
116
+ "transformer.blocks.25.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
117
+ "transformer.blocks.25.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
118
+ "transformer.blocks.25.norm_1.weight": "model-00002-of-00002.safetensors",
119
+ "transformer.blocks.25.norm_2.weight": "model-00002-of-00002.safetensors",
120
+ "transformer.blocks.26.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
121
+ "transformer.blocks.26.attn.out_proj.weight": "model-00002-of-00002.safetensors",
122
+ "transformer.blocks.26.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
123
+ "transformer.blocks.26.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
124
+ "transformer.blocks.26.norm_1.weight": "model-00002-of-00002.safetensors",
125
+ "transformer.blocks.26.norm_2.weight": "model-00002-of-00002.safetensors",
126
+ "transformer.blocks.27.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
127
+ "transformer.blocks.27.attn.out_proj.weight": "model-00002-of-00002.safetensors",
128
+ "transformer.blocks.27.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
129
+ "transformer.blocks.27.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
130
+ "transformer.blocks.27.norm_1.weight": "model-00002-of-00002.safetensors",
131
+ "transformer.blocks.27.norm_2.weight": "model-00002-of-00002.safetensors",
132
+ "transformer.blocks.28.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
133
+ "transformer.blocks.28.attn.out_proj.weight": "model-00002-of-00002.safetensors",
134
+ "transformer.blocks.28.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
135
+ "transformer.blocks.28.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
136
+ "transformer.blocks.28.norm_1.weight": "model-00002-of-00002.safetensors",
137
+ "transformer.blocks.28.norm_2.weight": "model-00002-of-00002.safetensors",
138
+ "transformer.blocks.29.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
139
+ "transformer.blocks.29.attn.out_proj.weight": "model-00002-of-00002.safetensors",
140
+ "transformer.blocks.29.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
141
+ "transformer.blocks.29.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
142
+ "transformer.blocks.29.norm_1.weight": "model-00002-of-00002.safetensors",
143
+ "transformer.blocks.29.norm_2.weight": "model-00002-of-00002.safetensors",
144
+ "transformer.blocks.3.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
145
+ "transformer.blocks.3.attn.out_proj.weight": "model-00001-of-00002.safetensors",
146
+ "transformer.blocks.3.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
147
+ "transformer.blocks.3.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
148
+ "transformer.blocks.3.norm_1.weight": "model-00001-of-00002.safetensors",
149
+ "transformer.blocks.3.norm_2.weight": "model-00001-of-00002.safetensors",
150
+ "transformer.blocks.30.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
151
+ "transformer.blocks.30.attn.out_proj.weight": "model-00002-of-00002.safetensors",
152
+ "transformer.blocks.30.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
153
+ "transformer.blocks.30.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
154
+ "transformer.blocks.30.norm_1.weight": "model-00002-of-00002.safetensors",
155
+ "transformer.blocks.30.norm_2.weight": "model-00002-of-00002.safetensors",
156
+ "transformer.blocks.31.attn.Wqkv.weight": "model-00002-of-00002.safetensors",
157
+ "transformer.blocks.31.attn.out_proj.weight": "model-00002-of-00002.safetensors",
158
+ "transformer.blocks.31.ffn.down_proj.weight": "model-00002-of-00002.safetensors",
159
+ "transformer.blocks.31.ffn.up_proj.weight": "model-00002-of-00002.safetensors",
160
+ "transformer.blocks.31.norm_1.weight": "model-00002-of-00002.safetensors",
161
+ "transformer.blocks.31.norm_2.weight": "model-00002-of-00002.safetensors",
162
+ "transformer.blocks.4.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
163
+ "transformer.blocks.4.attn.out_proj.weight": "model-00001-of-00002.safetensors",
164
+ "transformer.blocks.4.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
165
+ "transformer.blocks.4.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
166
+ "transformer.blocks.4.norm_1.weight": "model-00001-of-00002.safetensors",
167
+ "transformer.blocks.4.norm_2.weight": "model-00001-of-00002.safetensors",
168
+ "transformer.blocks.5.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
169
+ "transformer.blocks.5.attn.out_proj.weight": "model-00001-of-00002.safetensors",
170
+ "transformer.blocks.5.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
171
+ "transformer.blocks.5.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
172
+ "transformer.blocks.5.norm_1.weight": "model-00001-of-00002.safetensors",
173
+ "transformer.blocks.5.norm_2.weight": "model-00001-of-00002.safetensors",
174
+ "transformer.blocks.6.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
175
+ "transformer.blocks.6.attn.out_proj.weight": "model-00001-of-00002.safetensors",
176
+ "transformer.blocks.6.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
177
+ "transformer.blocks.6.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
178
+ "transformer.blocks.6.norm_1.weight": "model-00001-of-00002.safetensors",
179
+ "transformer.blocks.6.norm_2.weight": "model-00001-of-00002.safetensors",
180
+ "transformer.blocks.7.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
181
+ "transformer.blocks.7.attn.out_proj.weight": "model-00001-of-00002.safetensors",
182
+ "transformer.blocks.7.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
183
+ "transformer.blocks.7.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
184
+ "transformer.blocks.7.norm_1.weight": "model-00001-of-00002.safetensors",
185
+ "transformer.blocks.7.norm_2.weight": "model-00001-of-00002.safetensors",
186
+ "transformer.blocks.8.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
187
+ "transformer.blocks.8.attn.out_proj.weight": "model-00001-of-00002.safetensors",
188
+ "transformer.blocks.8.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
189
+ "transformer.blocks.8.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
190
+ "transformer.blocks.8.norm_1.weight": "model-00001-of-00002.safetensors",
191
+ "transformer.blocks.8.norm_2.weight": "model-00001-of-00002.safetensors",
192
+ "transformer.blocks.9.attn.Wqkv.weight": "model-00001-of-00002.safetensors",
193
+ "transformer.blocks.9.attn.out_proj.weight": "model-00001-of-00002.safetensors",
194
+ "transformer.blocks.9.ffn.down_proj.weight": "model-00001-of-00002.safetensors",
195
+ "transformer.blocks.9.ffn.up_proj.weight": "model-00001-of-00002.safetensors",
196
+ "transformer.blocks.9.norm_1.weight": "model-00001-of-00002.safetensors",
197
+ "transformer.blocks.9.norm_2.weight": "model-00001-of-00002.safetensors",
198
+ "transformer.norm_f.weight": "model-00002-of-00002.safetensors",
199
+ "transformer.wte.weight": "model-00001-of-00002.safetensors"
200
+ }
201
+ }
sft_output_simpleQA/checkpoint-2508/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2ba83ee07ad10e54499e2a884f0b1967a31dbbc0c9e5f43b5d41d1de6d93f9f
3
+ size 14748140538
sft_output_simpleQA/checkpoint-2508/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95b6047bd8cc6f4cdf7c46dea47edb8e542435510070c6cd1e0a7d9ccf5fd7da
3
+ size 14244
sft_output_simpleQA/checkpoint-2508/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb33588f8269547605dcf71e9d03456d8b26f2f7b5d096173f582d32bf3dd218
3
+ size 1064
sft_output_simpleQA/checkpoint-2508/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
sft_output_simpleQA/checkpoint-2508/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
sft_output_simpleQA/checkpoint-2508/tokenizer_config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<unk>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<s>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<pad>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ }
36
+ },
37
+ "bos_token": "<s>",
38
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{% if message['role'] == 'user' and loop.first %}{{ '### Câu hỏi: ' + message['content'].strip() }}{% elif message['role'] == 'user' %}{{ '\n### Câu hỏi: ' + message['content'].strip() }}{% elif message['role'] == 'assistant' %}{{ '\n### Trả lời: ' + message['content'] + eos_token }}{% endif %}{% if loop.last %}{% if message['role'] == 'user' and add_generation_prompt %}{{ '\n### Trả lời:' }}{% endif %}{% endif %}{% endfor %}",
39
+ "clean_up_tokenization_spaces": false,
40
+ "eos_token": "</s>",
41
+ "extra_special_tokens": {},
42
+ "merges_file": null,
43
+ "model_max_length": 8192,
44
+ "pad_token": "<pad>",
45
+ "padding_side": "right",
46
+ "tokenizer_class": "BloomTokenizer",
47
+ "unk_token": "<unk>",
48
+ "vocab_file": null
49
+ }
sft_output_simpleQA/checkpoint-2508/trainer_state.json ADDED
@@ -0,0 +1,2314 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 2508,
3
+ "best_metric": 7.677623271942139,
4
+ "best_model_checkpoint": "./sft_output_simpleQA/checkpoint-2508",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 2508,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.011961722488038277,
14
+ "grad_norm": 82.0,
15
+ "learning_rate": 4.982057416267943e-05,
16
+ "loss": 35.9519,
17
+ "mean_token_accuracy": 0.00015143249183893203,
18
+ "num_tokens": 16296.0,
19
+ "step": 10
20
+ },
21
+ {
22
+ "epoch": 0.023923444976076555,
23
+ "grad_norm": 88.0,
24
+ "learning_rate": 4.962121212121213e-05,
25
+ "loss": 34.0394,
26
+ "mean_token_accuracy": 0.00014820520300418139,
27
+ "num_tokens": 30078.0,
28
+ "step": 20
29
+ },
30
+ {
31
+ "epoch": 0.03588516746411483,
32
+ "grad_norm": 99.0,
33
+ "learning_rate": 4.9421850079744817e-05,
34
+ "loss": 32.9064,
35
+ "mean_token_accuracy": 0.0004058516991790384,
36
+ "num_tokens": 45602.0,
37
+ "step": 30
38
+ },
39
+ {
40
+ "epoch": 0.04784688995215311,
41
+ "grad_norm": 87.5,
42
+ "learning_rate": 4.922248803827751e-05,
43
+ "loss": 31.4831,
44
+ "mean_token_accuracy": 0.00040118659380823375,
45
+ "num_tokens": 58420.0,
46
+ "step": 40
47
+ },
48
+ {
49
+ "epoch": 0.05980861244019139,
50
+ "grad_norm": 83.0,
51
+ "learning_rate": 4.902312599681021e-05,
52
+ "loss": 29.9708,
53
+ "mean_token_accuracy": 0.00044854874722659587,
54
+ "num_tokens": 75126.0,
55
+ "step": 50
56
+ },
57
+ {
58
+ "epoch": 0.07177033492822966,
59
+ "grad_norm": 78.0,
60
+ "learning_rate": 4.8823763955342905e-05,
61
+ "loss": 28.5442,
62
+ "mean_token_accuracy": 0.00010988024296239018,
63
+ "num_tokens": 92306.0,
64
+ "step": 60
65
+ },
66
+ {
67
+ "epoch": 0.08373205741626795,
68
+ "grad_norm": 78.5,
69
+ "learning_rate": 4.86244019138756e-05,
70
+ "loss": 27.7258,
71
+ "mean_token_accuracy": 0.0008938259619753808,
72
+ "num_tokens": 108533.0,
73
+ "step": 70
74
+ },
75
+ {
76
+ "epoch": 0.09569377990430622,
77
+ "grad_norm": 75.5,
78
+ "learning_rate": 4.84250398724083e-05,
79
+ "loss": 26.4549,
80
+ "mean_token_accuracy": 0.0015082828351296485,
81
+ "num_tokens": 123975.0,
82
+ "step": 80
83
+ },
84
+ {
85
+ "epoch": 0.1076555023923445,
86
+ "grad_norm": 92.5,
87
+ "learning_rate": 4.8225677830940993e-05,
88
+ "loss": 25.4496,
89
+ "mean_token_accuracy": 0.0011613189650233834,
90
+ "num_tokens": 138192.0,
91
+ "step": 90
92
+ },
93
+ {
94
+ "epoch": 0.11961722488038277,
95
+ "grad_norm": 69.0,
96
+ "learning_rate": 4.802631578947368e-05,
97
+ "loss": 24.3682,
98
+ "mean_token_accuracy": 0.0030715125263668596,
99
+ "num_tokens": 155705.0,
100
+ "step": 100
101
+ },
102
+ {
103
+ "epoch": 0.13157894736842105,
104
+ "grad_norm": 68.5,
105
+ "learning_rate": 4.782695374800638e-05,
106
+ "loss": 22.8859,
107
+ "mean_token_accuracy": 0.0022452569246524945,
108
+ "num_tokens": 174596.0,
109
+ "step": 110
110
+ },
111
+ {
112
+ "epoch": 0.14354066985645933,
113
+ "grad_norm": 77.5,
114
+ "learning_rate": 4.7627591706539075e-05,
115
+ "loss": 22.18,
116
+ "mean_token_accuracy": 0.002278627915075049,
117
+ "num_tokens": 189101.0,
118
+ "step": 120
119
+ },
120
+ {
121
+ "epoch": 0.15550239234449761,
122
+ "grad_norm": 61.75,
123
+ "learning_rate": 4.742822966507177e-05,
124
+ "loss": 21.2046,
125
+ "mean_token_accuracy": 0.0015775824780575932,
126
+ "num_tokens": 205871.0,
127
+ "step": 130
128
+ },
129
+ {
130
+ "epoch": 0.1674641148325359,
131
+ "grad_norm": 60.25,
132
+ "learning_rate": 4.722886762360447e-05,
133
+ "loss": 20.4473,
134
+ "mean_token_accuracy": 0.0028540739440359175,
135
+ "num_tokens": 221139.0,
136
+ "step": 140
137
+ },
138
+ {
139
+ "epoch": 0.17942583732057416,
140
+ "grad_norm": 50.75,
141
+ "learning_rate": 4.7029505582137164e-05,
142
+ "loss": 19.1652,
143
+ "mean_token_accuracy": 0.0029712042480241506,
144
+ "num_tokens": 239226.0,
145
+ "step": 150
146
+ },
147
+ {
148
+ "epoch": 0.19138755980861244,
149
+ "grad_norm": 58.5,
150
+ "learning_rate": 4.683014354066986e-05,
151
+ "loss": 19.1199,
152
+ "mean_token_accuracy": 0.003232458623824641,
153
+ "num_tokens": 254905.0,
154
+ "step": 160
155
+ },
156
+ {
157
+ "epoch": 0.20334928229665072,
158
+ "grad_norm": 44.75,
159
+ "learning_rate": 4.6630781499202556e-05,
160
+ "loss": 18.0852,
161
+ "mean_token_accuracy": 0.0037016855902038516,
162
+ "num_tokens": 269854.0,
163
+ "step": 170
164
+ },
165
+ {
166
+ "epoch": 0.215311004784689,
167
+ "grad_norm": 47.25,
168
+ "learning_rate": 4.6431419457735245e-05,
169
+ "loss": 17.1532,
170
+ "mean_token_accuracy": 0.0039575109840370715,
171
+ "num_tokens": 285467.0,
172
+ "step": 180
173
+ },
174
+ {
175
+ "epoch": 0.22727272727272727,
176
+ "grad_norm": 46.25,
177
+ "learning_rate": 4.623205741626794e-05,
178
+ "loss": 16.3724,
179
+ "mean_token_accuracy": 0.005616716342046857,
180
+ "num_tokens": 300009.0,
181
+ "step": 190
182
+ },
183
+ {
184
+ "epoch": 0.23923444976076555,
185
+ "grad_norm": 36.25,
186
+ "learning_rate": 4.603269537480064e-05,
187
+ "loss": 15.551,
188
+ "mean_token_accuracy": 0.006238142051734031,
189
+ "num_tokens": 316750.0,
190
+ "step": 200
191
+ },
192
+ {
193
+ "epoch": 0.2511961722488038,
194
+ "grad_norm": 41.75,
195
+ "learning_rate": 4.5833333333333334e-05,
196
+ "loss": 14.3853,
197
+ "mean_token_accuracy": 0.007631447073072195,
198
+ "num_tokens": 332122.0,
199
+ "step": 210
200
+ },
201
+ {
202
+ "epoch": 0.2631578947368421,
203
+ "grad_norm": 35.25,
204
+ "learning_rate": 4.563397129186603e-05,
205
+ "loss": 14.1354,
206
+ "mean_token_accuracy": 0.011761691747233271,
207
+ "num_tokens": 345320.0,
208
+ "step": 220
209
+ },
210
+ {
211
+ "epoch": 0.2751196172248804,
212
+ "grad_norm": 45.25,
213
+ "learning_rate": 4.5434609250398726e-05,
214
+ "loss": 13.9972,
215
+ "mean_token_accuracy": 0.011121365474537015,
216
+ "num_tokens": 360052.0,
217
+ "step": 230
218
+ },
219
+ {
220
+ "epoch": 0.28708133971291866,
221
+ "grad_norm": 26.875,
222
+ "learning_rate": 4.523524720893142e-05,
223
+ "loss": 13.4802,
224
+ "mean_token_accuracy": 0.014945084601640702,
225
+ "num_tokens": 375674.0,
226
+ "step": 240
227
+ },
228
+ {
229
+ "epoch": 0.29904306220095694,
230
+ "grad_norm": 29.125,
231
+ "learning_rate": 4.503588516746411e-05,
232
+ "loss": 12.5142,
233
+ "mean_token_accuracy": 0.018064118921756744,
234
+ "num_tokens": 391484.0,
235
+ "step": 250
236
+ },
237
+ {
238
+ "epoch": 0.31100478468899523,
239
+ "grad_norm": 34.25,
240
+ "learning_rate": 4.483652312599681e-05,
241
+ "loss": 12.4949,
242
+ "mean_token_accuracy": 0.018949254974722864,
243
+ "num_tokens": 405586.0,
244
+ "step": 260
245
+ },
246
+ {
247
+ "epoch": 0.3229665071770335,
248
+ "grad_norm": 41.75,
249
+ "learning_rate": 4.4637161084529504e-05,
250
+ "loss": 12.1934,
251
+ "mean_token_accuracy": 0.017779586557298898,
252
+ "num_tokens": 421161.0,
253
+ "step": 270
254
+ },
255
+ {
256
+ "epoch": 0.3349282296650718,
257
+ "grad_norm": 29.625,
258
+ "learning_rate": 4.44377990430622e-05,
259
+ "loss": 11.9731,
260
+ "mean_token_accuracy": 0.019403476919978856,
261
+ "num_tokens": 438046.0,
262
+ "step": 280
263
+ },
264
+ {
265
+ "epoch": 0.34688995215311,
266
+ "grad_norm": 33.25,
267
+ "learning_rate": 4.4238437001594896e-05,
268
+ "loss": 11.2492,
269
+ "mean_token_accuracy": 0.021459076926112174,
270
+ "num_tokens": 454098.0,
271
+ "step": 290
272
+ },
273
+ {
274
+ "epoch": 0.3588516746411483,
275
+ "grad_norm": 25.875,
276
+ "learning_rate": 4.40390749601276e-05,
277
+ "loss": 11.2402,
278
+ "mean_token_accuracy": 0.0200654573738575,
279
+ "num_tokens": 471900.0,
280
+ "step": 300
281
+ },
282
+ {
283
+ "epoch": 0.3708133971291866,
284
+ "grad_norm": 29.375,
285
+ "learning_rate": 4.383971291866029e-05,
286
+ "loss": 11.1068,
287
+ "mean_token_accuracy": 0.022695281356573106,
288
+ "num_tokens": 486795.0,
289
+ "step": 310
290
+ },
291
+ {
292
+ "epoch": 0.3827751196172249,
293
+ "grad_norm": 32.25,
294
+ "learning_rate": 4.3640350877192985e-05,
295
+ "loss": 10.9675,
296
+ "mean_token_accuracy": 0.021525637619197367,
297
+ "num_tokens": 503071.0,
298
+ "step": 320
299
+ },
300
+ {
301
+ "epoch": 0.39473684210526316,
302
+ "grad_norm": 25.75,
303
+ "learning_rate": 4.344098883572568e-05,
304
+ "loss": 10.3584,
305
+ "mean_token_accuracy": 0.027966510131955148,
306
+ "num_tokens": 517177.0,
307
+ "step": 330
308
+ },
309
+ {
310
+ "epoch": 0.40669856459330145,
311
+ "grad_norm": 29.75,
312
+ "learning_rate": 4.324162679425838e-05,
313
+ "loss": 10.4161,
314
+ "mean_token_accuracy": 0.02753335628658533,
315
+ "num_tokens": 531444.0,
316
+ "step": 340
317
+ },
318
+ {
319
+ "epoch": 0.41866028708133973,
320
+ "grad_norm": 25.5,
321
+ "learning_rate": 4.304226475279107e-05,
322
+ "loss": 10.2745,
323
+ "mean_token_accuracy": 0.027429930120706558,
324
+ "num_tokens": 547939.0,
325
+ "step": 350
326
+ },
327
+ {
328
+ "epoch": 0.430622009569378,
329
+ "grad_norm": 39.5,
330
+ "learning_rate": 4.284290271132377e-05,
331
+ "loss": 10.5799,
332
+ "mean_token_accuracy": 0.026776039227843284,
333
+ "num_tokens": 563428.0,
334
+ "step": 360
335
+ },
336
+ {
337
+ "epoch": 0.44258373205741625,
338
+ "grad_norm": 21.375,
339
+ "learning_rate": 4.2643540669856466e-05,
340
+ "loss": 10.0957,
341
+ "mean_token_accuracy": 0.02906850315630436,
342
+ "num_tokens": 578486.0,
343
+ "step": 370
344
+ },
345
+ {
346
+ "epoch": 0.45454545454545453,
347
+ "grad_norm": 20.75,
348
+ "learning_rate": 4.244417862838916e-05,
349
+ "loss": 9.7364,
350
+ "mean_token_accuracy": 0.025704856589436532,
351
+ "num_tokens": 592561.0,
352
+ "step": 380
353
+ },
354
+ {
355
+ "epoch": 0.4665071770334928,
356
+ "grad_norm": 19.625,
357
+ "learning_rate": 4.224481658692185e-05,
358
+ "loss": 9.6673,
359
+ "mean_token_accuracy": 0.028244724869728087,
360
+ "num_tokens": 609128.0,
361
+ "step": 390
362
+ },
363
+ {
364
+ "epoch": 0.4784688995215311,
365
+ "grad_norm": 20.0,
366
+ "learning_rate": 4.204545454545455e-05,
367
+ "loss": 9.4029,
368
+ "mean_token_accuracy": 0.029186627827584743,
369
+ "num_tokens": 623498.0,
370
+ "step": 400
371
+ },
372
+ {
373
+ "epoch": 0.4904306220095694,
374
+ "grad_norm": 19.625,
375
+ "learning_rate": 4.1846092503987243e-05,
376
+ "loss": 9.6126,
377
+ "mean_token_accuracy": 0.02799250017851591,
378
+ "num_tokens": 639552.0,
379
+ "step": 410
380
+ },
381
+ {
382
+ "epoch": 0.5023923444976076,
383
+ "grad_norm": 18.5,
384
+ "learning_rate": 4.164673046251994e-05,
385
+ "loss": 9.4314,
386
+ "mean_token_accuracy": 0.029687324352562428,
387
+ "num_tokens": 654069.0,
388
+ "step": 420
389
+ },
390
+ {
391
+ "epoch": 0.5143540669856459,
392
+ "grad_norm": 21.875,
393
+ "learning_rate": 4.1447368421052636e-05,
394
+ "loss": 9.111,
395
+ "mean_token_accuracy": 0.028119230084121226,
396
+ "num_tokens": 669050.0,
397
+ "step": 430
398
+ },
399
+ {
400
+ "epoch": 0.5263157894736842,
401
+ "grad_norm": 16.5,
402
+ "learning_rate": 4.124800637958533e-05,
403
+ "loss": 9.6644,
404
+ "mean_token_accuracy": 0.03025283953174949,
405
+ "num_tokens": 684231.0,
406
+ "step": 440
407
+ },
408
+ {
409
+ "epoch": 0.5382775119617225,
410
+ "grad_norm": 16.5,
411
+ "learning_rate": 4.104864433811803e-05,
412
+ "loss": 9.3751,
413
+ "mean_token_accuracy": 0.02908811327069998,
414
+ "num_tokens": 702272.0,
415
+ "step": 450
416
+ },
417
+ {
418
+ "epoch": 0.5502392344497608,
419
+ "grad_norm": 17.375,
420
+ "learning_rate": 4.084928229665072e-05,
421
+ "loss": 9.1689,
422
+ "mean_token_accuracy": 0.030976119451224804,
423
+ "num_tokens": 720268.0,
424
+ "step": 460
425
+ },
426
+ {
427
+ "epoch": 0.562200956937799,
428
+ "grad_norm": 26.875,
429
+ "learning_rate": 4.0649920255183414e-05,
430
+ "loss": 9.0435,
431
+ "mean_token_accuracy": 0.03120351955294609,
432
+ "num_tokens": 738391.0,
433
+ "step": 470
434
+ },
435
+ {
436
+ "epoch": 0.5741626794258373,
437
+ "grad_norm": 16.125,
438
+ "learning_rate": 4.045055821371611e-05,
439
+ "loss": 8.8265,
440
+ "mean_token_accuracy": 0.03519163243472576,
441
+ "num_tokens": 752428.0,
442
+ "step": 480
443
+ },
444
+ {
445
+ "epoch": 0.5861244019138756,
446
+ "grad_norm": 16.75,
447
+ "learning_rate": 4.0251196172248806e-05,
448
+ "loss": 9.4865,
449
+ "mean_token_accuracy": 0.03259586580097675,
450
+ "num_tokens": 770127.0,
451
+ "step": 490
452
+ },
453
+ {
454
+ "epoch": 0.5980861244019139,
455
+ "grad_norm": 22.0,
456
+ "learning_rate": 4.00518341307815e-05,
457
+ "loss": 8.697,
458
+ "mean_token_accuracy": 0.03240090515464544,
459
+ "num_tokens": 786336.0,
460
+ "step": 500
461
+ },
462
+ {
463
+ "epoch": 0.6100478468899522,
464
+ "grad_norm": 21.5,
465
+ "learning_rate": 3.98524720893142e-05,
466
+ "loss": 8.8026,
467
+ "mean_token_accuracy": 0.028156513161957265,
468
+ "num_tokens": 802572.0,
469
+ "step": 510
470
+ },
471
+ {
472
+ "epoch": 0.6220095693779905,
473
+ "grad_norm": 14.5625,
474
+ "learning_rate": 3.9653110047846894e-05,
475
+ "loss": 8.5123,
476
+ "mean_token_accuracy": 0.03563346900045872,
477
+ "num_tokens": 817380.0,
478
+ "step": 520
479
+ },
480
+ {
481
+ "epoch": 0.6339712918660287,
482
+ "grad_norm": 19.125,
483
+ "learning_rate": 3.945374800637959e-05,
484
+ "loss": 8.5729,
485
+ "mean_token_accuracy": 0.03419111575931311,
486
+ "num_tokens": 834232.0,
487
+ "step": 530
488
+ },
489
+ {
490
+ "epoch": 0.645933014354067,
491
+ "grad_norm": 15.1875,
492
+ "learning_rate": 3.925438596491228e-05,
493
+ "loss": 8.9406,
494
+ "mean_token_accuracy": 0.03414075216278434,
495
+ "num_tokens": 850055.0,
496
+ "step": 540
497
+ },
498
+ {
499
+ "epoch": 0.6578947368421053,
500
+ "grad_norm": 25.875,
501
+ "learning_rate": 3.9055023923444976e-05,
502
+ "loss": 8.4975,
503
+ "mean_token_accuracy": 0.03595046941190958,
504
+ "num_tokens": 866135.0,
505
+ "step": 550
506
+ },
507
+ {
508
+ "epoch": 0.6698564593301436,
509
+ "grad_norm": 12.75,
510
+ "learning_rate": 3.885566188197767e-05,
511
+ "loss": 8.2385,
512
+ "mean_token_accuracy": 0.03583654277026653,
513
+ "num_tokens": 884639.0,
514
+ "step": 560
515
+ },
516
+ {
517
+ "epoch": 0.6818181818181818,
518
+ "grad_norm": 13.75,
519
+ "learning_rate": 3.865629984051037e-05,
520
+ "loss": 8.4336,
521
+ "mean_token_accuracy": 0.032993039861321447,
522
+ "num_tokens": 901515.0,
523
+ "step": 570
524
+ },
525
+ {
526
+ "epoch": 0.69377990430622,
527
+ "grad_norm": 21.0,
528
+ "learning_rate": 3.8456937799043065e-05,
529
+ "loss": 8.6048,
530
+ "mean_token_accuracy": 0.0326311569660902,
531
+ "num_tokens": 918500.0,
532
+ "step": 580
533
+ },
534
+ {
535
+ "epoch": 0.7057416267942583,
536
+ "grad_norm": 16.625,
537
+ "learning_rate": 3.825757575757576e-05,
538
+ "loss": 8.3489,
539
+ "mean_token_accuracy": 0.037175539322197436,
540
+ "num_tokens": 934654.0,
541
+ "step": 590
542
+ },
543
+ {
544
+ "epoch": 0.7177033492822966,
545
+ "grad_norm": 12.875,
546
+ "learning_rate": 3.805821371610846e-05,
547
+ "loss": 7.919,
548
+ "mean_token_accuracy": 0.040041149407625196,
549
+ "num_tokens": 951137.0,
550
+ "step": 600
551
+ },
552
+ {
553
+ "epoch": 0.7296650717703349,
554
+ "grad_norm": 20.0,
555
+ "learning_rate": 3.7858851674641146e-05,
556
+ "loss": 8.0379,
557
+ "mean_token_accuracy": 0.034196126647293566,
558
+ "num_tokens": 966104.0,
559
+ "step": 610
560
+ },
561
+ {
562
+ "epoch": 0.7416267942583732,
563
+ "grad_norm": 14.3125,
564
+ "learning_rate": 3.765948963317384e-05,
565
+ "loss": 8.185,
566
+ "mean_token_accuracy": 0.037530180625617506,
567
+ "num_tokens": 978723.0,
568
+ "step": 620
569
+ },
570
+ {
571
+ "epoch": 0.7535885167464115,
572
+ "grad_norm": 19.125,
573
+ "learning_rate": 3.746012759170654e-05,
574
+ "loss": 7.9819,
575
+ "mean_token_accuracy": 0.04008628278970718,
576
+ "num_tokens": 996286.0,
577
+ "step": 630
578
+ },
579
+ {
580
+ "epoch": 0.7655502392344498,
581
+ "grad_norm": 13.625,
582
+ "learning_rate": 3.7260765550239235e-05,
583
+ "loss": 7.7242,
584
+ "mean_token_accuracy": 0.03889909721910954,
585
+ "num_tokens": 1012265.0,
586
+ "step": 640
587
+ },
588
+ {
589
+ "epoch": 0.777511961722488,
590
+ "grad_norm": 19.375,
591
+ "learning_rate": 3.706140350877193e-05,
592
+ "loss": 7.9709,
593
+ "mean_token_accuracy": 0.04032406695187092,
594
+ "num_tokens": 1028334.0,
595
+ "step": 650
596
+ },
597
+ {
598
+ "epoch": 0.7894736842105263,
599
+ "grad_norm": 13.25,
600
+ "learning_rate": 3.686204146730463e-05,
601
+ "loss": 8.0423,
602
+ "mean_token_accuracy": 0.038657473400235176,
603
+ "num_tokens": 1043429.0,
604
+ "step": 660
605
+ },
606
+ {
607
+ "epoch": 0.8014354066985646,
608
+ "grad_norm": 14.375,
609
+ "learning_rate": 3.666267942583732e-05,
610
+ "loss": 7.9946,
611
+ "mean_token_accuracy": 0.03983959686011076,
612
+ "num_tokens": 1057274.0,
613
+ "step": 670
614
+ },
615
+ {
616
+ "epoch": 0.8133971291866029,
617
+ "grad_norm": 16.375,
618
+ "learning_rate": 3.646331738437001e-05,
619
+ "loss": 8.3528,
620
+ "mean_token_accuracy": 0.037851406075060365,
621
+ "num_tokens": 1071259.0,
622
+ "step": 680
623
+ },
624
+ {
625
+ "epoch": 0.8253588516746412,
626
+ "grad_norm": 15.3125,
627
+ "learning_rate": 3.626395534290271e-05,
628
+ "loss": 8.3224,
629
+ "mean_token_accuracy": 0.03589223362505436,
630
+ "num_tokens": 1089053.0,
631
+ "step": 690
632
+ },
633
+ {
634
+ "epoch": 0.8373205741626795,
635
+ "grad_norm": 20.375,
636
+ "learning_rate": 3.6064593301435405e-05,
637
+ "loss": 7.845,
638
+ "mean_token_accuracy": 0.03959129601716995,
639
+ "num_tokens": 1102285.0,
640
+ "step": 700
641
+ },
642
+ {
643
+ "epoch": 0.8492822966507177,
644
+ "grad_norm": 13.8125,
645
+ "learning_rate": 3.58652312599681e-05,
646
+ "loss": 7.8353,
647
+ "mean_token_accuracy": 0.046173612400889394,
648
+ "num_tokens": 1118501.0,
649
+ "step": 710
650
+ },
651
+ {
652
+ "epoch": 0.861244019138756,
653
+ "grad_norm": 14.625,
654
+ "learning_rate": 3.56658692185008e-05,
655
+ "loss": 8.0074,
656
+ "mean_token_accuracy": 0.03960017450153828,
657
+ "num_tokens": 1135070.0,
658
+ "step": 720
659
+ },
660
+ {
661
+ "epoch": 0.8732057416267942,
662
+ "grad_norm": 14.3125,
663
+ "learning_rate": 3.5466507177033493e-05,
664
+ "loss": 7.8224,
665
+ "mean_token_accuracy": 0.04144163206219673,
666
+ "num_tokens": 1150879.0,
667
+ "step": 730
668
+ },
669
+ {
670
+ "epoch": 0.8851674641148325,
671
+ "grad_norm": 15.25,
672
+ "learning_rate": 3.526714513556619e-05,
673
+ "loss": 7.8332,
674
+ "mean_token_accuracy": 0.04649510532617569,
675
+ "num_tokens": 1166334.0,
676
+ "step": 740
677
+ },
678
+ {
679
+ "epoch": 0.8971291866028708,
680
+ "grad_norm": 29.0,
681
+ "learning_rate": 3.5067783094098886e-05,
682
+ "loss": 8.3253,
683
+ "mean_token_accuracy": 0.03786750454455614,
684
+ "num_tokens": 1183891.0,
685
+ "step": 750
686
+ },
687
+ {
688
+ "epoch": 0.9090909090909091,
689
+ "grad_norm": 14.9375,
690
+ "learning_rate": 3.4868421052631575e-05,
691
+ "loss": 7.776,
692
+ "mean_token_accuracy": 0.04886236339807511,
693
+ "num_tokens": 1199155.0,
694
+ "step": 760
695
+ },
696
+ {
697
+ "epoch": 0.9210526315789473,
698
+ "grad_norm": 15.5,
699
+ "learning_rate": 3.466905901116427e-05,
700
+ "loss": 8.1109,
701
+ "mean_token_accuracy": 0.04119991734623909,
702
+ "num_tokens": 1212791.0,
703
+ "step": 770
704
+ },
705
+ {
706
+ "epoch": 0.9330143540669856,
707
+ "grad_norm": 13.75,
708
+ "learning_rate": 3.4469696969696974e-05,
709
+ "loss": 7.8783,
710
+ "mean_token_accuracy": 0.05077510289847851,
711
+ "num_tokens": 1227455.0,
712
+ "step": 780
713
+ },
714
+ {
715
+ "epoch": 0.9449760765550239,
716
+ "grad_norm": 23.875,
717
+ "learning_rate": 3.427033492822967e-05,
718
+ "loss": 8.0403,
719
+ "mean_token_accuracy": 0.041047198325395585,
720
+ "num_tokens": 1245814.0,
721
+ "step": 790
722
+ },
723
+ {
724
+ "epoch": 0.9569377990430622,
725
+ "grad_norm": 10.875,
726
+ "learning_rate": 3.4070972886762367e-05,
727
+ "loss": 7.9784,
728
+ "mean_token_accuracy": 0.044691753759980204,
729
+ "num_tokens": 1261146.0,
730
+ "step": 800
731
+ },
732
+ {
733
+ "epoch": 0.9688995215311005,
734
+ "grad_norm": 13.875,
735
+ "learning_rate": 3.387161084529506e-05,
736
+ "loss": 7.8961,
737
+ "mean_token_accuracy": 0.04515785053372383,
738
+ "num_tokens": 1279584.0,
739
+ "step": 810
740
+ },
741
+ {
742
+ "epoch": 0.9808612440191388,
743
+ "grad_norm": 9.25,
744
+ "learning_rate": 3.367224880382775e-05,
745
+ "loss": 7.6902,
746
+ "mean_token_accuracy": 0.04966730885207653,
747
+ "num_tokens": 1298622.0,
748
+ "step": 820
749
+ },
750
+ {
751
+ "epoch": 0.992822966507177,
752
+ "grad_norm": 10.625,
753
+ "learning_rate": 3.347288676236045e-05,
754
+ "loss": 7.6308,
755
+ "mean_token_accuracy": 0.051312201842665675,
756
+ "num_tokens": 1313079.0,
757
+ "step": 830
758
+ },
759
+ {
760
+ "epoch": 1.0,
761
+ "eval_loss": 7.889608860015869,
762
+ "eval_mean_token_accuracy": 0.04921758498223323,
763
+ "eval_num_tokens": 1321280.0,
764
+ "eval_runtime": 18.7031,
765
+ "eval_samples_per_second": 22.349,
766
+ "eval_steps_per_second": 2.834,
767
+ "step": 836
768
+ },
769
+ {
770
+ "epoch": 1.0047846889952152,
771
+ "grad_norm": 10.625,
772
+ "learning_rate": 3.3273524720893144e-05,
773
+ "loss": 7.6017,
774
+ "mean_token_accuracy": 0.04911007285118103,
775
+ "num_tokens": 1327489.0,
776
+ "step": 840
777
+ },
778
+ {
779
+ "epoch": 1.0167464114832536,
780
+ "grad_norm": 13.625,
781
+ "learning_rate": 3.307416267942584e-05,
782
+ "loss": 7.7674,
783
+ "mean_token_accuracy": 0.04897945895791054,
784
+ "num_tokens": 1343987.0,
785
+ "step": 850
786
+ },
787
+ {
788
+ "epoch": 1.0287081339712918,
789
+ "grad_norm": 15.625,
790
+ "learning_rate": 3.287480063795854e-05,
791
+ "loss": 7.6781,
792
+ "mean_token_accuracy": 0.04750379957258701,
793
+ "num_tokens": 1360126.0,
794
+ "step": 860
795
+ },
796
+ {
797
+ "epoch": 1.0406698564593302,
798
+ "grad_norm": 9.9375,
799
+ "learning_rate": 3.267543859649123e-05,
800
+ "loss": 7.4758,
801
+ "mean_token_accuracy": 0.04961836487054825,
802
+ "num_tokens": 1376926.0,
803
+ "step": 870
804
+ },
805
+ {
806
+ "epoch": 1.0526315789473684,
807
+ "grad_norm": 10.25,
808
+ "learning_rate": 3.247607655502393e-05,
809
+ "loss": 7.6265,
810
+ "mean_token_accuracy": 0.049552620574831965,
811
+ "num_tokens": 1392486.0,
812
+ "step": 880
813
+ },
814
+ {
815
+ "epoch": 1.0645933014354068,
816
+ "grad_norm": 11.5625,
817
+ "learning_rate": 3.227671451355662e-05,
818
+ "loss": 8.0101,
819
+ "mean_token_accuracy": 0.04870691932737827,
820
+ "num_tokens": 1409932.0,
821
+ "step": 890
822
+ },
823
+ {
824
+ "epoch": 1.076555023923445,
825
+ "grad_norm": 14.375,
826
+ "learning_rate": 3.2077352472089315e-05,
827
+ "loss": 7.6739,
828
+ "mean_token_accuracy": 0.048780468106269834,
829
+ "num_tokens": 1427292.0,
830
+ "step": 900
831
+ },
832
+ {
833
+ "epoch": 1.0885167464114833,
834
+ "grad_norm": 10.625,
835
+ "learning_rate": 3.187799043062201e-05,
836
+ "loss": 7.7819,
837
+ "mean_token_accuracy": 0.0481975007802248,
838
+ "num_tokens": 1443541.0,
839
+ "step": 910
840
+ },
841
+ {
842
+ "epoch": 1.1004784688995215,
843
+ "grad_norm": 40.75,
844
+ "learning_rate": 3.167862838915471e-05,
845
+ "loss": 7.7404,
846
+ "mean_token_accuracy": 0.048559962399303915,
847
+ "num_tokens": 1457736.0,
848
+ "step": 920
849
+ },
850
+ {
851
+ "epoch": 1.11244019138756,
852
+ "grad_norm": 14.4375,
853
+ "learning_rate": 3.14792663476874e-05,
854
+ "loss": 7.8908,
855
+ "mean_token_accuracy": 0.04830890707671642,
856
+ "num_tokens": 1472783.0,
857
+ "step": 930
858
+ },
859
+ {
860
+ "epoch": 1.124401913875598,
861
+ "grad_norm": 13.0,
862
+ "learning_rate": 3.12799043062201e-05,
863
+ "loss": 7.562,
864
+ "mean_token_accuracy": 0.052664054557681084,
865
+ "num_tokens": 1489101.0,
866
+ "step": 940
867
+ },
868
+ {
869
+ "epoch": 1.1363636363636362,
870
+ "grad_norm": 9.625,
871
+ "learning_rate": 3.1080542264752795e-05,
872
+ "loss": 7.5301,
873
+ "mean_token_accuracy": 0.05596833899617195,
874
+ "num_tokens": 1505652.0,
875
+ "step": 950
876
+ },
877
+ {
878
+ "epoch": 1.1483253588516746,
879
+ "grad_norm": 12.3125,
880
+ "learning_rate": 3.088118022328549e-05,
881
+ "loss": 7.9597,
882
+ "mean_token_accuracy": 0.051548514328897,
883
+ "num_tokens": 1520909.0,
884
+ "step": 960
885
+ },
886
+ {
887
+ "epoch": 1.160287081339713,
888
+ "grad_norm": 13.3125,
889
+ "learning_rate": 3.068181818181818e-05,
890
+ "loss": 7.649,
891
+ "mean_token_accuracy": 0.052403326518833636,
892
+ "num_tokens": 1535652.0,
893
+ "step": 970
894
+ },
895
+ {
896
+ "epoch": 1.1722488038277512,
897
+ "grad_norm": 13.375,
898
+ "learning_rate": 3.048245614035088e-05,
899
+ "loss": 7.6625,
900
+ "mean_token_accuracy": 0.048475722596049306,
901
+ "num_tokens": 1552205.0,
902
+ "step": 980
903
+ },
904
+ {
905
+ "epoch": 1.1842105263157894,
906
+ "grad_norm": 10.3125,
907
+ "learning_rate": 3.0283094098883573e-05,
908
+ "loss": 7.3846,
909
+ "mean_token_accuracy": 0.053659194707870485,
910
+ "num_tokens": 1567310.0,
911
+ "step": 990
912
+ },
913
+ {
914
+ "epoch": 1.1961722488038278,
915
+ "grad_norm": 13.125,
916
+ "learning_rate": 3.008373205741627e-05,
917
+ "loss": 7.6728,
918
+ "mean_token_accuracy": 0.062439772114157674,
919
+ "num_tokens": 1581582.0,
920
+ "step": 1000
921
+ },
922
+ {
923
+ "epoch": 1.208133971291866,
924
+ "grad_norm": 13.25,
925
+ "learning_rate": 2.9884370015948966e-05,
926
+ "loss": 7.619,
927
+ "mean_token_accuracy": 0.04972767271101475,
928
+ "num_tokens": 1598082.0,
929
+ "step": 1010
930
+ },
931
+ {
932
+ "epoch": 1.2200956937799043,
933
+ "grad_norm": 11.25,
934
+ "learning_rate": 2.968500797448166e-05,
935
+ "loss": 7.681,
936
+ "mean_token_accuracy": 0.04762520119547844,
937
+ "num_tokens": 1613460.0,
938
+ "step": 1020
939
+ },
940
+ {
941
+ "epoch": 1.2320574162679425,
942
+ "grad_norm": 15.3125,
943
+ "learning_rate": 2.9485645933014355e-05,
944
+ "loss": 7.9789,
945
+ "mean_token_accuracy": 0.053393319994211194,
946
+ "num_tokens": 1631405.0,
947
+ "step": 1030
948
+ },
949
+ {
950
+ "epoch": 1.244019138755981,
951
+ "grad_norm": 13.875,
952
+ "learning_rate": 2.928628389154705e-05,
953
+ "loss": 7.5258,
954
+ "mean_token_accuracy": 0.0497327771037817,
955
+ "num_tokens": 1646315.0,
956
+ "step": 1040
957
+ },
958
+ {
959
+ "epoch": 1.255980861244019,
960
+ "grad_norm": 13.625,
961
+ "learning_rate": 2.9086921850079747e-05,
962
+ "loss": 7.5095,
963
+ "mean_token_accuracy": 0.04922787994146347,
964
+ "num_tokens": 1663451.0,
965
+ "step": 1050
966
+ },
967
+ {
968
+ "epoch": 1.2679425837320575,
969
+ "grad_norm": 15.5,
970
+ "learning_rate": 2.888755980861244e-05,
971
+ "loss": 7.7784,
972
+ "mean_token_accuracy": 0.055617674812674524,
973
+ "num_tokens": 1677447.0,
974
+ "step": 1060
975
+ },
976
+ {
977
+ "epoch": 1.2799043062200957,
978
+ "grad_norm": 11.9375,
979
+ "learning_rate": 2.8688197767145136e-05,
980
+ "loss": 7.5099,
981
+ "mean_token_accuracy": 0.04883741177618504,
982
+ "num_tokens": 1695963.0,
983
+ "step": 1070
984
+ },
985
+ {
986
+ "epoch": 1.291866028708134,
987
+ "grad_norm": 11.0,
988
+ "learning_rate": 2.8488835725677832e-05,
989
+ "loss": 7.8607,
990
+ "mean_token_accuracy": 0.04243286289274693,
991
+ "num_tokens": 1713184.0,
992
+ "step": 1080
993
+ },
994
+ {
995
+ "epoch": 1.3038277511961722,
996
+ "grad_norm": 11.875,
997
+ "learning_rate": 2.8289473684210528e-05,
998
+ "loss": 7.6724,
999
+ "mean_token_accuracy": 0.05059143900871277,
1000
+ "num_tokens": 1727104.0,
1001
+ "step": 1090
1002
+ },
1003
+ {
1004
+ "epoch": 1.3157894736842106,
1005
+ "grad_norm": 10.5625,
1006
+ "learning_rate": 2.809011164274322e-05,
1007
+ "loss": 7.7283,
1008
+ "mean_token_accuracy": 0.05400256849825382,
1009
+ "num_tokens": 1739981.0,
1010
+ "step": 1100
1011
+ },
1012
+ {
1013
+ "epoch": 1.3277511961722488,
1014
+ "grad_norm": 10.625,
1015
+ "learning_rate": 2.7890749601275917e-05,
1016
+ "loss": 7.6728,
1017
+ "mean_token_accuracy": 0.05072182305157184,
1018
+ "num_tokens": 1756367.0,
1019
+ "step": 1110
1020
+ },
1021
+ {
1022
+ "epoch": 1.339712918660287,
1023
+ "grad_norm": 15.25,
1024
+ "learning_rate": 2.7691387559808613e-05,
1025
+ "loss": 7.5879,
1026
+ "mean_token_accuracy": 0.0517768245190382,
1027
+ "num_tokens": 1770201.0,
1028
+ "step": 1120
1029
+ },
1030
+ {
1031
+ "epoch": 1.3516746411483254,
1032
+ "grad_norm": 28.25,
1033
+ "learning_rate": 2.749202551834131e-05,
1034
+ "loss": 8.4438,
1035
+ "mean_token_accuracy": 0.054370812699198724,
1036
+ "num_tokens": 1784812.0,
1037
+ "step": 1130
1038
+ },
1039
+ {
1040
+ "epoch": 1.3636363636363638,
1041
+ "grad_norm": 13.0625,
1042
+ "learning_rate": 2.7292663476874002e-05,
1043
+ "loss": 7.7018,
1044
+ "mean_token_accuracy": 0.05111252702772617,
1045
+ "num_tokens": 1800658.0,
1046
+ "step": 1140
1047
+ },
1048
+ {
1049
+ "epoch": 1.375598086124402,
1050
+ "grad_norm": 20.125,
1051
+ "learning_rate": 2.7093301435406698e-05,
1052
+ "loss": 7.7442,
1053
+ "mean_token_accuracy": 0.050131586566567424,
1054
+ "num_tokens": 1817111.0,
1055
+ "step": 1150
1056
+ },
1057
+ {
1058
+ "epoch": 1.38755980861244,
1059
+ "grad_norm": 15.375,
1060
+ "learning_rate": 2.6893939393939394e-05,
1061
+ "loss": 7.8474,
1062
+ "mean_token_accuracy": 0.051092632487416266,
1063
+ "num_tokens": 1832422.0,
1064
+ "step": 1160
1065
+ },
1066
+ {
1067
+ "epoch": 1.3995215311004785,
1068
+ "grad_norm": 13.625,
1069
+ "learning_rate": 2.6694577352472087e-05,
1070
+ "loss": 7.8175,
1071
+ "mean_token_accuracy": 0.04797025993466377,
1072
+ "num_tokens": 1849302.0,
1073
+ "step": 1170
1074
+ },
1075
+ {
1076
+ "epoch": 1.4114832535885167,
1077
+ "grad_norm": 9.9375,
1078
+ "learning_rate": 2.6495215311004783e-05,
1079
+ "loss": 7.7398,
1080
+ "mean_token_accuracy": 0.05692324116826057,
1081
+ "num_tokens": 1863943.0,
1082
+ "step": 1180
1083
+ },
1084
+ {
1085
+ "epoch": 1.423444976076555,
1086
+ "grad_norm": 10.875,
1087
+ "learning_rate": 2.629585326953748e-05,
1088
+ "loss": 7.5425,
1089
+ "mean_token_accuracy": 0.055062346160411835,
1090
+ "num_tokens": 1878423.0,
1091
+ "step": 1190
1092
+ },
1093
+ {
1094
+ "epoch": 1.4354066985645932,
1095
+ "grad_norm": 13.125,
1096
+ "learning_rate": 2.6096491228070176e-05,
1097
+ "loss": 7.7075,
1098
+ "mean_token_accuracy": 0.05213572233915329,
1099
+ "num_tokens": 1893221.0,
1100
+ "step": 1200
1101
+ },
1102
+ {
1103
+ "epoch": 1.4473684210526316,
1104
+ "grad_norm": 10.6875,
1105
+ "learning_rate": 2.589712918660287e-05,
1106
+ "loss": 7.6832,
1107
+ "mean_token_accuracy": 0.05459029637277126,
1108
+ "num_tokens": 1910194.0,
1109
+ "step": 1210
1110
+ },
1111
+ {
1112
+ "epoch": 1.4593301435406698,
1113
+ "grad_norm": 16.25,
1114
+ "learning_rate": 2.5697767145135565e-05,
1115
+ "loss": 7.5755,
1116
+ "mean_token_accuracy": 0.05433895252645016,
1117
+ "num_tokens": 1927848.0,
1118
+ "step": 1220
1119
+ },
1120
+ {
1121
+ "epoch": 1.4712918660287082,
1122
+ "grad_norm": 11.1875,
1123
+ "learning_rate": 2.549840510366826e-05,
1124
+ "loss": 7.61,
1125
+ "mean_token_accuracy": 0.05629041828215122,
1126
+ "num_tokens": 1947519.0,
1127
+ "step": 1230
1128
+ },
1129
+ {
1130
+ "epoch": 1.4832535885167464,
1131
+ "grad_norm": 13.8125,
1132
+ "learning_rate": 2.5299043062200957e-05,
1133
+ "loss": 7.6037,
1134
+ "mean_token_accuracy": 0.057362372800707814,
1135
+ "num_tokens": 1962606.0,
1136
+ "step": 1240
1137
+ },
1138
+ {
1139
+ "epoch": 1.4952153110047846,
1140
+ "grad_norm": 12.4375,
1141
+ "learning_rate": 2.509968102073365e-05,
1142
+ "loss": 7.9281,
1143
+ "mean_token_accuracy": 0.05096495188772678,
1144
+ "num_tokens": 1979220.0,
1145
+ "step": 1250
1146
+ },
1147
+ {
1148
+ "epoch": 1.507177033492823,
1149
+ "grad_norm": 11.3125,
1150
+ "learning_rate": 2.490031897926635e-05,
1151
+ "loss": 7.5714,
1152
+ "mean_token_accuracy": 0.051651550084352495,
1153
+ "num_tokens": 1995455.0,
1154
+ "step": 1260
1155
+ },
1156
+ {
1157
+ "epoch": 1.5191387559808613,
1158
+ "grad_norm": 16.625,
1159
+ "learning_rate": 2.4700956937799045e-05,
1160
+ "loss": 7.8689,
1161
+ "mean_token_accuracy": 0.04886260032653809,
1162
+ "num_tokens": 2012448.0,
1163
+ "step": 1270
1164
+ },
1165
+ {
1166
+ "epoch": 1.5311004784688995,
1167
+ "grad_norm": 12.3125,
1168
+ "learning_rate": 2.4501594896331738e-05,
1169
+ "loss": 7.5645,
1170
+ "mean_token_accuracy": 0.05832184366881847,
1171
+ "num_tokens": 2029242.0,
1172
+ "step": 1280
1173
+ },
1174
+ {
1175
+ "epoch": 1.5430622009569377,
1176
+ "grad_norm": 9.0625,
1177
+ "learning_rate": 2.4302232854864434e-05,
1178
+ "loss": 7.6672,
1179
+ "mean_token_accuracy": 0.05073030404746533,
1180
+ "num_tokens": 2045026.0,
1181
+ "step": 1290
1182
+ },
1183
+ {
1184
+ "epoch": 1.555023923444976,
1185
+ "grad_norm": 16.125,
1186
+ "learning_rate": 2.410287081339713e-05,
1187
+ "loss": 8.02,
1188
+ "mean_token_accuracy": 0.05017628595232963,
1189
+ "num_tokens": 2059452.0,
1190
+ "step": 1300
1191
+ },
1192
+ {
1193
+ "epoch": 1.5669856459330145,
1194
+ "grad_norm": 13.375,
1195
+ "learning_rate": 2.3903508771929827e-05,
1196
+ "loss": 7.5553,
1197
+ "mean_token_accuracy": 0.05098568722605705,
1198
+ "num_tokens": 2076536.0,
1199
+ "step": 1310
1200
+ },
1201
+ {
1202
+ "epoch": 1.5789473684210527,
1203
+ "grad_norm": 14.875,
1204
+ "learning_rate": 2.370414673046252e-05,
1205
+ "loss": 7.6946,
1206
+ "mean_token_accuracy": 0.053875621780753134,
1207
+ "num_tokens": 2092058.0,
1208
+ "step": 1320
1209
+ },
1210
+ {
1211
+ "epoch": 1.5909090909090908,
1212
+ "grad_norm": 10.5625,
1213
+ "learning_rate": 2.3504784688995216e-05,
1214
+ "loss": 7.815,
1215
+ "mean_token_accuracy": 0.05061427466571331,
1216
+ "num_tokens": 2108006.0,
1217
+ "step": 1330
1218
+ },
1219
+ {
1220
+ "epoch": 1.6028708133971292,
1221
+ "grad_norm": 16.0,
1222
+ "learning_rate": 2.3305422647527912e-05,
1223
+ "loss": 7.8098,
1224
+ "mean_token_accuracy": 0.048553376272320745,
1225
+ "num_tokens": 2125780.0,
1226
+ "step": 1340
1227
+ },
1228
+ {
1229
+ "epoch": 1.6148325358851676,
1230
+ "grad_norm": 13.375,
1231
+ "learning_rate": 2.3106060606060605e-05,
1232
+ "loss": 7.4606,
1233
+ "mean_token_accuracy": 0.06270428001880646,
1234
+ "num_tokens": 2140186.0,
1235
+ "step": 1350
1236
+ },
1237
+ {
1238
+ "epoch": 1.6267942583732058,
1239
+ "grad_norm": 13.1875,
1240
+ "learning_rate": 2.29066985645933e-05,
1241
+ "loss": 7.5675,
1242
+ "mean_token_accuracy": 0.05657584816217422,
1243
+ "num_tokens": 2157025.0,
1244
+ "step": 1360
1245
+ },
1246
+ {
1247
+ "epoch": 1.638755980861244,
1248
+ "grad_norm": 12.5,
1249
+ "learning_rate": 2.2707336523125997e-05,
1250
+ "loss": 7.4467,
1251
+ "mean_token_accuracy": 0.053974258899688723,
1252
+ "num_tokens": 2171122.0,
1253
+ "step": 1370
1254
+ },
1255
+ {
1256
+ "epoch": 1.6507177033492821,
1257
+ "grad_norm": 10.875,
1258
+ "learning_rate": 2.2507974481658693e-05,
1259
+ "loss": 7.4626,
1260
+ "mean_token_accuracy": 0.052587350085377696,
1261
+ "num_tokens": 2186298.0,
1262
+ "step": 1380
1263
+ },
1264
+ {
1265
+ "epoch": 1.6626794258373205,
1266
+ "grad_norm": 12.9375,
1267
+ "learning_rate": 2.230861244019139e-05,
1268
+ "loss": 7.5117,
1269
+ "mean_token_accuracy": 0.055986383184790614,
1270
+ "num_tokens": 2201725.0,
1271
+ "step": 1390
1272
+ },
1273
+ {
1274
+ "epoch": 1.674641148325359,
1275
+ "grad_norm": 14.4375,
1276
+ "learning_rate": 2.2109250398724085e-05,
1277
+ "loss": 7.7233,
1278
+ "mean_token_accuracy": 0.05318257175385952,
1279
+ "num_tokens": 2216843.0,
1280
+ "step": 1400
1281
+ },
1282
+ {
1283
+ "epoch": 1.686602870813397,
1284
+ "grad_norm": 16.75,
1285
+ "learning_rate": 2.190988835725678e-05,
1286
+ "loss": 7.6867,
1287
+ "mean_token_accuracy": 0.052845988795161246,
1288
+ "num_tokens": 2234658.0,
1289
+ "step": 1410
1290
+ },
1291
+ {
1292
+ "epoch": 1.6985645933014353,
1293
+ "grad_norm": 9.8125,
1294
+ "learning_rate": 2.1710526315789474e-05,
1295
+ "loss": 7.5536,
1296
+ "mean_token_accuracy": 0.05676522888243198,
1297
+ "num_tokens": 2249972.0,
1298
+ "step": 1420
1299
+ },
1300
+ {
1301
+ "epoch": 1.7105263157894737,
1302
+ "grad_norm": 15.25,
1303
+ "learning_rate": 2.151116427432217e-05,
1304
+ "loss": 7.8057,
1305
+ "mean_token_accuracy": 0.052781854569911954,
1306
+ "num_tokens": 2265621.0,
1307
+ "step": 1430
1308
+ },
1309
+ {
1310
+ "epoch": 1.722488038277512,
1311
+ "grad_norm": 12.125,
1312
+ "learning_rate": 2.1311802232854867e-05,
1313
+ "loss": 7.7877,
1314
+ "mean_token_accuracy": 0.05187459997832775,
1315
+ "num_tokens": 2281490.0,
1316
+ "step": 1440
1317
+ },
1318
+ {
1319
+ "epoch": 1.7344497607655502,
1320
+ "grad_norm": 14.25,
1321
+ "learning_rate": 2.1112440191387563e-05,
1322
+ "loss": 7.5293,
1323
+ "mean_token_accuracy": 0.05277816876769066,
1324
+ "num_tokens": 2296372.0,
1325
+ "step": 1450
1326
+ },
1327
+ {
1328
+ "epoch": 1.7464114832535884,
1329
+ "grad_norm": 10.75,
1330
+ "learning_rate": 2.0913078149920256e-05,
1331
+ "loss": 7.5161,
1332
+ "mean_token_accuracy": 0.059847253188490865,
1333
+ "num_tokens": 2311710.0,
1334
+ "step": 1460
1335
+ },
1336
+ {
1337
+ "epoch": 1.7583732057416268,
1338
+ "grad_norm": 17.125,
1339
+ "learning_rate": 2.071371610845295e-05,
1340
+ "loss": 7.7205,
1341
+ "mean_token_accuracy": 0.05046251565217972,
1342
+ "num_tokens": 2328125.0,
1343
+ "step": 1470
1344
+ },
1345
+ {
1346
+ "epoch": 1.7703349282296652,
1347
+ "grad_norm": 15.1875,
1348
+ "learning_rate": 2.0514354066985648e-05,
1349
+ "loss": 7.4838,
1350
+ "mean_token_accuracy": 0.05662818215787411,
1351
+ "num_tokens": 2342723.0,
1352
+ "step": 1480
1353
+ },
1354
+ {
1355
+ "epoch": 1.7822966507177034,
1356
+ "grad_norm": 15.6875,
1357
+ "learning_rate": 2.0314992025518344e-05,
1358
+ "loss": 7.5797,
1359
+ "mean_token_accuracy": 0.05622794292867184,
1360
+ "num_tokens": 2357907.0,
1361
+ "step": 1490
1362
+ },
1363
+ {
1364
+ "epoch": 1.7942583732057416,
1365
+ "grad_norm": 19.75,
1366
+ "learning_rate": 2.0115629984051037e-05,
1367
+ "loss": 7.8304,
1368
+ "mean_token_accuracy": 0.05124356746673584,
1369
+ "num_tokens": 2374672.0,
1370
+ "step": 1500
1371
+ },
1372
+ {
1373
+ "epoch": 1.80622009569378,
1374
+ "grad_norm": 10.75,
1375
+ "learning_rate": 1.9916267942583733e-05,
1376
+ "loss": 7.6402,
1377
+ "mean_token_accuracy": 0.05051091276109219,
1378
+ "num_tokens": 2391528.0,
1379
+ "step": 1510
1380
+ },
1381
+ {
1382
+ "epoch": 1.8181818181818183,
1383
+ "grad_norm": 15.125,
1384
+ "learning_rate": 1.971690590111643e-05,
1385
+ "loss": 7.4673,
1386
+ "mean_token_accuracy": 0.05444490201771259,
1387
+ "num_tokens": 2407044.0,
1388
+ "step": 1520
1389
+ },
1390
+ {
1391
+ "epoch": 1.8301435406698565,
1392
+ "grad_norm": 10.8125,
1393
+ "learning_rate": 1.9517543859649122e-05,
1394
+ "loss": 7.647,
1395
+ "mean_token_accuracy": 0.048004547134041786,
1396
+ "num_tokens": 2422247.0,
1397
+ "step": 1530
1398
+ },
1399
+ {
1400
+ "epoch": 1.8421052631578947,
1401
+ "grad_norm": 17.375,
1402
+ "learning_rate": 1.9318181818181818e-05,
1403
+ "loss": 7.582,
1404
+ "mean_token_accuracy": 0.05293577797710895,
1405
+ "num_tokens": 2438387.0,
1406
+ "step": 1540
1407
+ },
1408
+ {
1409
+ "epoch": 1.8540669856459329,
1410
+ "grad_norm": 14.9375,
1411
+ "learning_rate": 1.9118819776714514e-05,
1412
+ "loss": 7.6128,
1413
+ "mean_token_accuracy": 0.05296882390975952,
1414
+ "num_tokens": 2454353.0,
1415
+ "step": 1550
1416
+ },
1417
+ {
1418
+ "epoch": 1.8660287081339713,
1419
+ "grad_norm": 14.8125,
1420
+ "learning_rate": 1.891945773524721e-05,
1421
+ "loss": 7.5914,
1422
+ "mean_token_accuracy": 0.05177001543343067,
1423
+ "num_tokens": 2469893.0,
1424
+ "step": 1560
1425
+ },
1426
+ {
1427
+ "epoch": 1.8779904306220097,
1428
+ "grad_norm": 12.9375,
1429
+ "learning_rate": 1.8720095693779903e-05,
1430
+ "loss": 7.8388,
1431
+ "mean_token_accuracy": 0.05344080198556185,
1432
+ "num_tokens": 2485080.0,
1433
+ "step": 1570
1434
+ },
1435
+ {
1436
+ "epoch": 1.8899521531100478,
1437
+ "grad_norm": 13.375,
1438
+ "learning_rate": 1.85207336523126e-05,
1439
+ "loss": 7.4122,
1440
+ "mean_token_accuracy": 0.056378521770238874,
1441
+ "num_tokens": 2499300.0,
1442
+ "step": 1580
1443
+ },
1444
+ {
1445
+ "epoch": 1.901913875598086,
1446
+ "grad_norm": 9.4375,
1447
+ "learning_rate": 1.8321371610845295e-05,
1448
+ "loss": 7.4838,
1449
+ "mean_token_accuracy": 0.052063405886292456,
1450
+ "num_tokens": 2515672.0,
1451
+ "step": 1590
1452
+ },
1453
+ {
1454
+ "epoch": 1.9138755980861244,
1455
+ "grad_norm": 10.25,
1456
+ "learning_rate": 1.812200956937799e-05,
1457
+ "loss": 7.6252,
1458
+ "mean_token_accuracy": 0.052615062519907954,
1459
+ "num_tokens": 2532131.0,
1460
+ "step": 1600
1461
+ },
1462
+ {
1463
+ "epoch": 1.9258373205741628,
1464
+ "grad_norm": 13.9375,
1465
+ "learning_rate": 1.7922647527910684e-05,
1466
+ "loss": 7.4704,
1467
+ "mean_token_accuracy": 0.05482649989426136,
1468
+ "num_tokens": 2548166.0,
1469
+ "step": 1610
1470
+ },
1471
+ {
1472
+ "epoch": 1.937799043062201,
1473
+ "grad_norm": 11.125,
1474
+ "learning_rate": 1.7723285486443384e-05,
1475
+ "loss": 7.5689,
1476
+ "mean_token_accuracy": 0.05435803644359112,
1477
+ "num_tokens": 2562836.0,
1478
+ "step": 1620
1479
+ },
1480
+ {
1481
+ "epoch": 1.9497607655502391,
1482
+ "grad_norm": 12.0,
1483
+ "learning_rate": 1.752392344497608e-05,
1484
+ "loss": 7.652,
1485
+ "mean_token_accuracy": 0.05930311307311058,
1486
+ "num_tokens": 2576424.0,
1487
+ "step": 1630
1488
+ },
1489
+ {
1490
+ "epoch": 1.9617224880382775,
1491
+ "grad_norm": 12.0,
1492
+ "learning_rate": 1.7324561403508773e-05,
1493
+ "loss": 7.5536,
1494
+ "mean_token_accuracy": 0.05557992160320282,
1495
+ "num_tokens": 2592898.0,
1496
+ "step": 1640
1497
+ },
1498
+ {
1499
+ "epoch": 1.973684210526316,
1500
+ "grad_norm": 11.9375,
1501
+ "learning_rate": 1.712519936204147e-05,
1502
+ "loss": 8.0379,
1503
+ "mean_token_accuracy": 0.05598388984799385,
1504
+ "num_tokens": 2608168.0,
1505
+ "step": 1650
1506
+ },
1507
+ {
1508
+ "epoch": 1.985645933014354,
1509
+ "grad_norm": 12.75,
1510
+ "learning_rate": 1.6925837320574165e-05,
1511
+ "loss": 7.5948,
1512
+ "mean_token_accuracy": 0.0535994254052639,
1513
+ "num_tokens": 2624729.0,
1514
+ "step": 1660
1515
+ },
1516
+ {
1517
+ "epoch": 1.9976076555023923,
1518
+ "grad_norm": 17.875,
1519
+ "learning_rate": 1.6726475279106858e-05,
1520
+ "loss": 7.5617,
1521
+ "mean_token_accuracy": 0.055696948617696765,
1522
+ "num_tokens": 2640184.0,
1523
+ "step": 1670
1524
+ },
1525
+ {
1526
+ "epoch": 2.0,
1527
+ "eval_loss": 7.681546211242676,
1528
+ "eval_mean_token_accuracy": 0.05555312321433481,
1529
+ "eval_num_tokens": 2642560.0,
1530
+ "eval_runtime": 18.6343,
1531
+ "eval_samples_per_second": 22.432,
1532
+ "eval_steps_per_second": 2.844,
1533
+ "step": 1672
1534
+ },
1535
+ {
1536
+ "epoch": 2.0095693779904304,
1537
+ "grad_norm": 14.0,
1538
+ "learning_rate": 1.6527113237639554e-05,
1539
+ "loss": 7.7797,
1540
+ "mean_token_accuracy": 0.06362233757972717,
1541
+ "num_tokens": 2654711.0,
1542
+ "step": 1680
1543
+ },
1544
+ {
1545
+ "epoch": 2.021531100478469,
1546
+ "grad_norm": 10.0625,
1547
+ "learning_rate": 1.632775119617225e-05,
1548
+ "loss": 7.9745,
1549
+ "mean_token_accuracy": 0.057430309057235715,
1550
+ "num_tokens": 2669744.0,
1551
+ "step": 1690
1552
+ },
1553
+ {
1554
+ "epoch": 2.0334928229665072,
1555
+ "grad_norm": 12.5625,
1556
+ "learning_rate": 1.6128389154704946e-05,
1557
+ "loss": 7.5017,
1558
+ "mean_token_accuracy": 0.05466706827282906,
1559
+ "num_tokens": 2687783.0,
1560
+ "step": 1700
1561
+ },
1562
+ {
1563
+ "epoch": 2.0454545454545454,
1564
+ "grad_norm": 12.75,
1565
+ "learning_rate": 1.592902711323764e-05,
1566
+ "loss": 7.5026,
1567
+ "mean_token_accuracy": 0.053967179358005525,
1568
+ "num_tokens": 2705217.0,
1569
+ "step": 1710
1570
+ },
1571
+ {
1572
+ "epoch": 2.0574162679425836,
1573
+ "grad_norm": 13.9375,
1574
+ "learning_rate": 1.5729665071770335e-05,
1575
+ "loss": 7.6858,
1576
+ "mean_token_accuracy": 0.04923243746161461,
1577
+ "num_tokens": 2719654.0,
1578
+ "step": 1720
1579
+ },
1580
+ {
1581
+ "epoch": 2.069377990430622,
1582
+ "grad_norm": 17.5,
1583
+ "learning_rate": 1.553030303030303e-05,
1584
+ "loss": 7.6658,
1585
+ "mean_token_accuracy": 0.05302671529352665,
1586
+ "num_tokens": 2736205.0,
1587
+ "step": 1730
1588
+ },
1589
+ {
1590
+ "epoch": 2.0813397129186604,
1591
+ "grad_norm": 10.75,
1592
+ "learning_rate": 1.5330940988835728e-05,
1593
+ "loss": 8.0743,
1594
+ "mean_token_accuracy": 0.05337436944246292,
1595
+ "num_tokens": 2750160.0,
1596
+ "step": 1740
1597
+ },
1598
+ {
1599
+ "epoch": 2.0933014354066986,
1600
+ "grad_norm": 18.125,
1601
+ "learning_rate": 1.5131578947368422e-05,
1602
+ "loss": 7.9227,
1603
+ "mean_token_accuracy": 0.052077919244766235,
1604
+ "num_tokens": 2765861.0,
1605
+ "step": 1750
1606
+ },
1607
+ {
1608
+ "epoch": 2.1052631578947367,
1609
+ "grad_norm": 11.1875,
1610
+ "learning_rate": 1.4932216905901117e-05,
1611
+ "loss": 7.6672,
1612
+ "mean_token_accuracy": 0.05186783894896507,
1613
+ "num_tokens": 2781742.0,
1614
+ "step": 1760
1615
+ },
1616
+ {
1617
+ "epoch": 2.117224880382775,
1618
+ "grad_norm": 10.3125,
1619
+ "learning_rate": 1.4732854864433813e-05,
1620
+ "loss": 7.7002,
1621
+ "mean_token_accuracy": 0.05188237279653549,
1622
+ "num_tokens": 2799667.0,
1623
+ "step": 1770
1624
+ },
1625
+ {
1626
+ "epoch": 2.1291866028708135,
1627
+ "grad_norm": 13.5625,
1628
+ "learning_rate": 1.4533492822966507e-05,
1629
+ "loss": 7.4252,
1630
+ "mean_token_accuracy": 0.053877873718738555,
1631
+ "num_tokens": 2816009.0,
1632
+ "step": 1780
1633
+ },
1634
+ {
1635
+ "epoch": 2.1411483253588517,
1636
+ "grad_norm": 17.125,
1637
+ "learning_rate": 1.4334130781499203e-05,
1638
+ "loss": 7.535,
1639
+ "mean_token_accuracy": 0.05251451525837183,
1640
+ "num_tokens": 2833484.0,
1641
+ "step": 1790
1642
+ },
1643
+ {
1644
+ "epoch": 2.15311004784689,
1645
+ "grad_norm": 11.875,
1646
+ "learning_rate": 1.4134768740031898e-05,
1647
+ "loss": 7.4562,
1648
+ "mean_token_accuracy": 0.054729873687028883,
1649
+ "num_tokens": 2848560.0,
1650
+ "step": 1800
1651
+ },
1652
+ {
1653
+ "epoch": 2.165071770334928,
1654
+ "grad_norm": 10.4375,
1655
+ "learning_rate": 1.3935406698564592e-05,
1656
+ "loss": 7.3528,
1657
+ "mean_token_accuracy": 0.05922457054257393,
1658
+ "num_tokens": 2861737.0,
1659
+ "step": 1810
1660
+ },
1661
+ {
1662
+ "epoch": 2.1770334928229667,
1663
+ "grad_norm": 13.75,
1664
+ "learning_rate": 1.3736044657097288e-05,
1665
+ "loss": 7.515,
1666
+ "mean_token_accuracy": 0.056258665025234224,
1667
+ "num_tokens": 2876906.0,
1668
+ "step": 1820
1669
+ },
1670
+ {
1671
+ "epoch": 2.188995215311005,
1672
+ "grad_norm": 13.375,
1673
+ "learning_rate": 1.3536682615629983e-05,
1674
+ "loss": 7.662,
1675
+ "mean_token_accuracy": 0.05119372643530369,
1676
+ "num_tokens": 2893621.0,
1677
+ "step": 1830
1678
+ },
1679
+ {
1680
+ "epoch": 2.200956937799043,
1681
+ "grad_norm": 10.5,
1682
+ "learning_rate": 1.3337320574162679e-05,
1683
+ "loss": 7.5924,
1684
+ "mean_token_accuracy": 0.0524879340082407,
1685
+ "num_tokens": 2910153.0,
1686
+ "step": 1840
1687
+ },
1688
+ {
1689
+ "epoch": 2.212918660287081,
1690
+ "grad_norm": 9.375,
1691
+ "learning_rate": 1.3137958532695374e-05,
1692
+ "loss": 7.5488,
1693
+ "mean_token_accuracy": 0.05508296974003315,
1694
+ "num_tokens": 2925796.0,
1695
+ "step": 1850
1696
+ },
1697
+ {
1698
+ "epoch": 2.22488038277512,
1699
+ "grad_norm": 15.75,
1700
+ "learning_rate": 1.2938596491228071e-05,
1701
+ "loss": 7.6105,
1702
+ "mean_token_accuracy": 0.05919160135090351,
1703
+ "num_tokens": 2944005.0,
1704
+ "step": 1860
1705
+ },
1706
+ {
1707
+ "epoch": 2.236842105263158,
1708
+ "grad_norm": 13.5625,
1709
+ "learning_rate": 1.2739234449760768e-05,
1710
+ "loss": 7.3525,
1711
+ "mean_token_accuracy": 0.057943170145154,
1712
+ "num_tokens": 2959480.0,
1713
+ "step": 1870
1714
+ },
1715
+ {
1716
+ "epoch": 2.248803827751196,
1717
+ "grad_norm": 10.1875,
1718
+ "learning_rate": 1.2539872408293462e-05,
1719
+ "loss": 7.7431,
1720
+ "mean_token_accuracy": 0.06011910997331142,
1721
+ "num_tokens": 2975510.0,
1722
+ "step": 1880
1723
+ },
1724
+ {
1725
+ "epoch": 2.2607655502392343,
1726
+ "grad_norm": 9.375,
1727
+ "learning_rate": 1.2340510366826157e-05,
1728
+ "loss": 7.4682,
1729
+ "mean_token_accuracy": 0.05141804702579975,
1730
+ "num_tokens": 2993262.0,
1731
+ "step": 1890
1732
+ },
1733
+ {
1734
+ "epoch": 2.2727272727272725,
1735
+ "grad_norm": 23.125,
1736
+ "learning_rate": 1.2141148325358851e-05,
1737
+ "loss": 7.7091,
1738
+ "mean_token_accuracy": 0.05523250661790371,
1739
+ "num_tokens": 3008482.0,
1740
+ "step": 1900
1741
+ },
1742
+ {
1743
+ "epoch": 2.284688995215311,
1744
+ "grad_norm": 12.0625,
1745
+ "learning_rate": 1.1941786283891547e-05,
1746
+ "loss": 7.6048,
1747
+ "mean_token_accuracy": 0.05504187680780888,
1748
+ "num_tokens": 3023533.0,
1749
+ "step": 1910
1750
+ },
1751
+ {
1752
+ "epoch": 2.2966507177033493,
1753
+ "grad_norm": 11.875,
1754
+ "learning_rate": 1.1742424242424243e-05,
1755
+ "loss": 7.4626,
1756
+ "mean_token_accuracy": 0.05014841184020043,
1757
+ "num_tokens": 3039230.0,
1758
+ "step": 1920
1759
+ },
1760
+ {
1761
+ "epoch": 2.3086124401913874,
1762
+ "grad_norm": 13.3125,
1763
+ "learning_rate": 1.154306220095694e-05,
1764
+ "loss": 7.4728,
1765
+ "mean_token_accuracy": 0.051592012494802476,
1766
+ "num_tokens": 3055324.0,
1767
+ "step": 1930
1768
+ },
1769
+ {
1770
+ "epoch": 2.320574162679426,
1771
+ "grad_norm": 14.1875,
1772
+ "learning_rate": 1.1343700159489634e-05,
1773
+ "loss": 7.6298,
1774
+ "mean_token_accuracy": 0.051040302589535715,
1775
+ "num_tokens": 3071645.0,
1776
+ "step": 1940
1777
+ },
1778
+ {
1779
+ "epoch": 2.3325358851674642,
1780
+ "grad_norm": 16.625,
1781
+ "learning_rate": 1.114433811802233e-05,
1782
+ "loss": 7.6678,
1783
+ "mean_token_accuracy": 0.05561281181871891,
1784
+ "num_tokens": 3085756.0,
1785
+ "step": 1950
1786
+ },
1787
+ {
1788
+ "epoch": 2.3444976076555024,
1789
+ "grad_norm": 14.625,
1790
+ "learning_rate": 1.0944976076555025e-05,
1791
+ "loss": 7.5681,
1792
+ "mean_token_accuracy": 0.05392424911260605,
1793
+ "num_tokens": 3100796.0,
1794
+ "step": 1960
1795
+ },
1796
+ {
1797
+ "epoch": 2.3564593301435406,
1798
+ "grad_norm": 11.625,
1799
+ "learning_rate": 1.074561403508772e-05,
1800
+ "loss": 7.5295,
1801
+ "mean_token_accuracy": 0.05229526199400425,
1802
+ "num_tokens": 3115388.0,
1803
+ "step": 1970
1804
+ },
1805
+ {
1806
+ "epoch": 2.3684210526315788,
1807
+ "grad_norm": 11.625,
1808
+ "learning_rate": 1.0546251993620415e-05,
1809
+ "loss": 7.4864,
1810
+ "mean_token_accuracy": 0.055051923543214795,
1811
+ "num_tokens": 3130295.0,
1812
+ "step": 1980
1813
+ },
1814
+ {
1815
+ "epoch": 2.3803827751196174,
1816
+ "grad_norm": 10.9375,
1817
+ "learning_rate": 1.034688995215311e-05,
1818
+ "loss": 7.8901,
1819
+ "mean_token_accuracy": 0.057028749212622645,
1820
+ "num_tokens": 3145780.0,
1821
+ "step": 1990
1822
+ },
1823
+ {
1824
+ "epoch": 2.3923444976076556,
1825
+ "grad_norm": 13.5625,
1826
+ "learning_rate": 1.0147527910685806e-05,
1827
+ "loss": 7.3899,
1828
+ "mean_token_accuracy": 0.05296322256326676,
1829
+ "num_tokens": 3162641.0,
1830
+ "step": 2000
1831
+ },
1832
+ {
1833
+ "epoch": 2.4043062200956937,
1834
+ "grad_norm": 12.4375,
1835
+ "learning_rate": 9.9481658692185e-06,
1836
+ "loss": 7.4156,
1837
+ "mean_token_accuracy": 0.05216502845287323,
1838
+ "num_tokens": 3180101.0,
1839
+ "step": 2010
1840
+ },
1841
+ {
1842
+ "epoch": 2.416267942583732,
1843
+ "grad_norm": 9.4375,
1844
+ "learning_rate": 9.748803827751196e-06,
1845
+ "loss": 7.6813,
1846
+ "mean_token_accuracy": 0.05489225313067436,
1847
+ "num_tokens": 3195441.0,
1848
+ "step": 2020
1849
+ },
1850
+ {
1851
+ "epoch": 2.4282296650717705,
1852
+ "grad_norm": 14.125,
1853
+ "learning_rate": 9.549441786283891e-06,
1854
+ "loss": 7.5118,
1855
+ "mean_token_accuracy": 0.05190498977899551,
1856
+ "num_tokens": 3209632.0,
1857
+ "step": 2030
1858
+ },
1859
+ {
1860
+ "epoch": 2.4401913875598087,
1861
+ "grad_norm": 12.0625,
1862
+ "learning_rate": 9.350079744816589e-06,
1863
+ "loss": 7.4348,
1864
+ "mean_token_accuracy": 0.05217030458152294,
1865
+ "num_tokens": 3227006.0,
1866
+ "step": 2040
1867
+ },
1868
+ {
1869
+ "epoch": 2.452153110047847,
1870
+ "grad_norm": 19.75,
1871
+ "learning_rate": 9.150717703349283e-06,
1872
+ "loss": 7.7842,
1873
+ "mean_token_accuracy": 0.05061689857393503,
1874
+ "num_tokens": 3243617.0,
1875
+ "step": 2050
1876
+ },
1877
+ {
1878
+ "epoch": 2.464114832535885,
1879
+ "grad_norm": 16.5,
1880
+ "learning_rate": 8.95135566188198e-06,
1881
+ "loss": 7.8997,
1882
+ "mean_token_accuracy": 0.05135177038609982,
1883
+ "num_tokens": 3262345.0,
1884
+ "step": 2060
1885
+ },
1886
+ {
1887
+ "epoch": 2.4760765550239237,
1888
+ "grad_norm": 12.4375,
1889
+ "learning_rate": 8.751993620414674e-06,
1890
+ "loss": 7.5425,
1891
+ "mean_token_accuracy": 0.05693710744380951,
1892
+ "num_tokens": 3280049.0,
1893
+ "step": 2070
1894
+ },
1895
+ {
1896
+ "epoch": 2.488038277511962,
1897
+ "grad_norm": 12.125,
1898
+ "learning_rate": 8.552631578947368e-06,
1899
+ "loss": 7.7683,
1900
+ "mean_token_accuracy": 0.052999266609549525,
1901
+ "num_tokens": 3293627.0,
1902
+ "step": 2080
1903
+ },
1904
+ {
1905
+ "epoch": 2.5,
1906
+ "grad_norm": 18.5,
1907
+ "learning_rate": 8.353269537480064e-06,
1908
+ "loss": 7.5021,
1909
+ "mean_token_accuracy": 0.0573310736566782,
1910
+ "num_tokens": 3308963.0,
1911
+ "step": 2090
1912
+ },
1913
+ {
1914
+ "epoch": 2.511961722488038,
1915
+ "grad_norm": 11.1875,
1916
+ "learning_rate": 8.153907496012759e-06,
1917
+ "loss": 7.5266,
1918
+ "mean_token_accuracy": 0.055393803864717484,
1919
+ "num_tokens": 3327211.0,
1920
+ "step": 2100
1921
+ },
1922
+ {
1923
+ "epoch": 2.5239234449760763,
1924
+ "grad_norm": 12.4375,
1925
+ "learning_rate": 7.954545454545455e-06,
1926
+ "loss": 8.0202,
1927
+ "mean_token_accuracy": 0.0516157153993845,
1928
+ "num_tokens": 3342704.0,
1929
+ "step": 2110
1930
+ },
1931
+ {
1932
+ "epoch": 2.535885167464115,
1933
+ "grad_norm": 11.75,
1934
+ "learning_rate": 7.75518341307815e-06,
1935
+ "loss": 7.6579,
1936
+ "mean_token_accuracy": 0.053000685200095175,
1937
+ "num_tokens": 3356957.0,
1938
+ "step": 2120
1939
+ },
1940
+ {
1941
+ "epoch": 2.547846889952153,
1942
+ "grad_norm": 12.0,
1943
+ "learning_rate": 7.555821371610846e-06,
1944
+ "loss": 7.5,
1945
+ "mean_token_accuracy": 0.05583856776356697,
1946
+ "num_tokens": 3371814.0,
1947
+ "step": 2130
1948
+ },
1949
+ {
1950
+ "epoch": 2.5598086124401913,
1951
+ "grad_norm": 11.25,
1952
+ "learning_rate": 7.35645933014354e-06,
1953
+ "loss": 7.4175,
1954
+ "mean_token_accuracy": 0.056200267374515535,
1955
+ "num_tokens": 3387835.0,
1956
+ "step": 2140
1957
+ },
1958
+ {
1959
+ "epoch": 2.57177033492823,
1960
+ "grad_norm": 14.75,
1961
+ "learning_rate": 7.1570972886762355e-06,
1962
+ "loss": 7.601,
1963
+ "mean_token_accuracy": 0.058812175691127774,
1964
+ "num_tokens": 3402724.0,
1965
+ "step": 2150
1966
+ },
1967
+ {
1968
+ "epoch": 2.583732057416268,
1969
+ "grad_norm": 15.125,
1970
+ "learning_rate": 6.9577352472089325e-06,
1971
+ "loss": 7.6731,
1972
+ "mean_token_accuracy": 0.05726858489215374,
1973
+ "num_tokens": 3415782.0,
1974
+ "step": 2160
1975
+ },
1976
+ {
1977
+ "epoch": 2.5956937799043063,
1978
+ "grad_norm": 16.25,
1979
+ "learning_rate": 6.758373205741628e-06,
1980
+ "loss": 7.9197,
1981
+ "mean_token_accuracy": 0.04841956198215484,
1982
+ "num_tokens": 3430907.0,
1983
+ "step": 2170
1984
+ },
1985
+ {
1986
+ "epoch": 2.6076555023923444,
1987
+ "grad_norm": 15.125,
1988
+ "learning_rate": 6.559011164274323e-06,
1989
+ "loss": 7.4308,
1990
+ "mean_token_accuracy": 0.05858968757092953,
1991
+ "num_tokens": 3445725.0,
1992
+ "step": 2180
1993
+ },
1994
+ {
1995
+ "epoch": 2.6196172248803826,
1996
+ "grad_norm": 13.5,
1997
+ "learning_rate": 6.3596491228070184e-06,
1998
+ "loss": 7.7449,
1999
+ "mean_token_accuracy": 0.05409410893917084,
2000
+ "num_tokens": 3463327.0,
2001
+ "step": 2190
2002
+ },
2003
+ {
2004
+ "epoch": 2.6315789473684212,
2005
+ "grad_norm": 12.125,
2006
+ "learning_rate": 6.160287081339714e-06,
2007
+ "loss": 7.5525,
2008
+ "mean_token_accuracy": 0.0532541248947382,
2009
+ "num_tokens": 3477759.0,
2010
+ "step": 2200
2011
+ },
2012
+ {
2013
+ "epoch": 2.6435406698564594,
2014
+ "grad_norm": 9.6875,
2015
+ "learning_rate": 5.960925039872409e-06,
2016
+ "loss": 7.408,
2017
+ "mean_token_accuracy": 0.06432338766753673,
2018
+ "num_tokens": 3493604.0,
2019
+ "step": 2210
2020
+ },
2021
+ {
2022
+ "epoch": 2.6555023923444976,
2023
+ "grad_norm": 18.25,
2024
+ "learning_rate": 5.7615629984051035e-06,
2025
+ "loss": 8.0218,
2026
+ "mean_token_accuracy": 0.053415299765765666,
2027
+ "num_tokens": 3508326.0,
2028
+ "step": 2220
2029
+ },
2030
+ {
2031
+ "epoch": 2.6674641148325358,
2032
+ "grad_norm": 11.25,
2033
+ "learning_rate": 5.562200956937799e-06,
2034
+ "loss": 7.7336,
2035
+ "mean_token_accuracy": 0.053274786472320555,
2036
+ "num_tokens": 3523788.0,
2037
+ "step": 2230
2038
+ },
2039
+ {
2040
+ "epoch": 2.679425837320574,
2041
+ "grad_norm": 9.75,
2042
+ "learning_rate": 5.362838915470494e-06,
2043
+ "loss": 7.5168,
2044
+ "mean_token_accuracy": 0.06089800409972668,
2045
+ "num_tokens": 3537455.0,
2046
+ "step": 2240
2047
+ },
2048
+ {
2049
+ "epoch": 2.6913875598086126,
2050
+ "grad_norm": 9.0,
2051
+ "learning_rate": 5.16347687400319e-06,
2052
+ "loss": 7.9244,
2053
+ "mean_token_accuracy": 0.05054541490972042,
2054
+ "num_tokens": 3554477.0,
2055
+ "step": 2250
2056
+ },
2057
+ {
2058
+ "epoch": 2.7033492822966507,
2059
+ "grad_norm": 14.0625,
2060
+ "learning_rate": 4.964114832535886e-06,
2061
+ "loss": 7.493,
2062
+ "mean_token_accuracy": 0.05583830550312996,
2063
+ "num_tokens": 3568763.0,
2064
+ "step": 2260
2065
+ },
2066
+ {
2067
+ "epoch": 2.715311004784689,
2068
+ "grad_norm": 10.6875,
2069
+ "learning_rate": 4.764752791068581e-06,
2070
+ "loss": 7.6837,
2071
+ "mean_token_accuracy": 0.056372621655464174,
2072
+ "num_tokens": 3584199.0,
2073
+ "step": 2270
2074
+ },
2075
+ {
2076
+ "epoch": 2.7272727272727275,
2077
+ "grad_norm": 8.875,
2078
+ "learning_rate": 4.565390749601276e-06,
2079
+ "loss": 7.5824,
2080
+ "mean_token_accuracy": 0.05312733016908169,
2081
+ "num_tokens": 3601486.0,
2082
+ "step": 2280
2083
+ },
2084
+ {
2085
+ "epoch": 2.7392344497607657,
2086
+ "grad_norm": 13.0,
2087
+ "learning_rate": 4.3660287081339716e-06,
2088
+ "loss": 7.5518,
2089
+ "mean_token_accuracy": 0.05006753616034985,
2090
+ "num_tokens": 3620786.0,
2091
+ "step": 2290
2092
+ },
2093
+ {
2094
+ "epoch": 2.751196172248804,
2095
+ "grad_norm": 13.875,
2096
+ "learning_rate": 4.166666666666667e-06,
2097
+ "loss": 7.3837,
2098
+ "mean_token_accuracy": 0.059043745324015615,
2099
+ "num_tokens": 3636570.0,
2100
+ "step": 2300
2101
+ },
2102
+ {
2103
+ "epoch": 2.763157894736842,
2104
+ "grad_norm": 12.5,
2105
+ "learning_rate": 3.967304625199362e-06,
2106
+ "loss": 7.523,
2107
+ "mean_token_accuracy": 0.05304129309952259,
2108
+ "num_tokens": 3651913.0,
2109
+ "step": 2310
2110
+ },
2111
+ {
2112
+ "epoch": 2.77511961722488,
2113
+ "grad_norm": 11.875,
2114
+ "learning_rate": 3.767942583732058e-06,
2115
+ "loss": 7.5162,
2116
+ "mean_token_accuracy": 0.055253866314888,
2117
+ "num_tokens": 3668785.0,
2118
+ "step": 2320
2119
+ },
2120
+ {
2121
+ "epoch": 2.787081339712919,
2122
+ "grad_norm": 16.625,
2123
+ "learning_rate": 3.5685805422647532e-06,
2124
+ "loss": 7.3978,
2125
+ "mean_token_accuracy": 0.05370482616126537,
2126
+ "num_tokens": 3685088.0,
2127
+ "step": 2330
2128
+ },
2129
+ {
2130
+ "epoch": 2.799043062200957,
2131
+ "grad_norm": 49.25,
2132
+ "learning_rate": 3.3692185007974485e-06,
2133
+ "loss": 7.9911,
2134
+ "mean_token_accuracy": 0.05654057189822197,
2135
+ "num_tokens": 3699717.0,
2136
+ "step": 2340
2137
+ },
2138
+ {
2139
+ "epoch": 2.811004784688995,
2140
+ "grad_norm": 9.1875,
2141
+ "learning_rate": 3.1698564593301434e-06,
2142
+ "loss": 7.3809,
2143
+ "mean_token_accuracy": 0.05578604489564896,
2144
+ "num_tokens": 3712729.0,
2145
+ "step": 2350
2146
+ },
2147
+ {
2148
+ "epoch": 2.8229665071770333,
2149
+ "grad_norm": 12.0625,
2150
+ "learning_rate": 2.970494417862839e-06,
2151
+ "loss": 7.4346,
2152
+ "mean_token_accuracy": 0.054243154078722,
2153
+ "num_tokens": 3727180.0,
2154
+ "step": 2360
2155
+ },
2156
+ {
2157
+ "epoch": 2.8349282296650715,
2158
+ "grad_norm": 15.1875,
2159
+ "learning_rate": 2.7711323763955345e-06,
2160
+ "loss": 7.4536,
2161
+ "mean_token_accuracy": 0.05370322801172733,
2162
+ "num_tokens": 3743937.0,
2163
+ "step": 2370
2164
+ },
2165
+ {
2166
+ "epoch": 2.84688995215311,
2167
+ "grad_norm": 27.875,
2168
+ "learning_rate": 2.57177033492823e-06,
2169
+ "loss": 7.5259,
2170
+ "mean_token_accuracy": 0.05500344783067703,
2171
+ "num_tokens": 3760306.0,
2172
+ "step": 2380
2173
+ },
2174
+ {
2175
+ "epoch": 2.8588516746411483,
2176
+ "grad_norm": 11.5,
2177
+ "learning_rate": 2.372408293460925e-06,
2178
+ "loss": 7.7337,
2179
+ "mean_token_accuracy": 0.05269717015326023,
2180
+ "num_tokens": 3776435.0,
2181
+ "step": 2390
2182
+ },
2183
+ {
2184
+ "epoch": 2.8708133971291865,
2185
+ "grad_norm": 11.75,
2186
+ "learning_rate": 2.1730462519936204e-06,
2187
+ "loss": 7.6882,
2188
+ "mean_token_accuracy": 0.06056210286915302,
2189
+ "num_tokens": 3792698.0,
2190
+ "step": 2400
2191
+ },
2192
+ {
2193
+ "epoch": 2.882775119617225,
2194
+ "grad_norm": 14.375,
2195
+ "learning_rate": 1.9736842105263157e-06,
2196
+ "loss": 7.3884,
2197
+ "mean_token_accuracy": 0.058717917650938034,
2198
+ "num_tokens": 3808389.0,
2199
+ "step": 2410
2200
+ },
2201
+ {
2202
+ "epoch": 2.8947368421052633,
2203
+ "grad_norm": 18.5,
2204
+ "learning_rate": 1.7743221690590115e-06,
2205
+ "loss": 7.7266,
2206
+ "mean_token_accuracy": 0.05015675462782383,
2207
+ "num_tokens": 3826096.0,
2208
+ "step": 2420
2209
+ },
2210
+ {
2211
+ "epoch": 2.9066985645933014,
2212
+ "grad_norm": 9.6875,
2213
+ "learning_rate": 1.5749601275917066e-06,
2214
+ "loss": 7.5727,
2215
+ "mean_token_accuracy": 0.055614912137389186,
2216
+ "num_tokens": 3841264.0,
2217
+ "step": 2430
2218
+ },
2219
+ {
2220
+ "epoch": 2.9186602870813396,
2221
+ "grad_norm": 15.75,
2222
+ "learning_rate": 1.375598086124402e-06,
2223
+ "loss": 7.7109,
2224
+ "mean_token_accuracy": 0.054452039673924445,
2225
+ "num_tokens": 3858262.0,
2226
+ "step": 2440
2227
+ },
2228
+ {
2229
+ "epoch": 2.930622009569378,
2230
+ "grad_norm": 13.5,
2231
+ "learning_rate": 1.1762360446570972e-06,
2232
+ "loss": 7.4082,
2233
+ "mean_token_accuracy": 0.05675496719777584,
2234
+ "num_tokens": 3874012.0,
2235
+ "step": 2450
2236
+ },
2237
+ {
2238
+ "epoch": 2.9425837320574164,
2239
+ "grad_norm": 11.9375,
2240
+ "learning_rate": 9.768740031897927e-07,
2241
+ "loss": 7.4929,
2242
+ "mean_token_accuracy": 0.05380352921783924,
2243
+ "num_tokens": 3889829.0,
2244
+ "step": 2460
2245
+ },
2246
+ {
2247
+ "epoch": 2.9545454545454546,
2248
+ "grad_norm": 13.4375,
2249
+ "learning_rate": 7.77511961722488e-07,
2250
+ "loss": 7.4973,
2251
+ "mean_token_accuracy": 0.057183431833982466,
2252
+ "num_tokens": 3905985.0,
2253
+ "step": 2470
2254
+ },
2255
+ {
2256
+ "epoch": 2.9665071770334928,
2257
+ "grad_norm": 23.875,
2258
+ "learning_rate": 5.781499202551834e-07,
2259
+ "loss": 7.7017,
2260
+ "mean_token_accuracy": 0.05155903100967407,
2261
+ "num_tokens": 3921607.0,
2262
+ "step": 2480
2263
+ },
2264
+ {
2265
+ "epoch": 2.9784688995215314,
2266
+ "grad_norm": 11.8125,
2267
+ "learning_rate": 3.787878787878788e-07,
2268
+ "loss": 7.7229,
2269
+ "mean_token_accuracy": 0.05097230635583401,
2270
+ "num_tokens": 3937739.0,
2271
+ "step": 2490
2272
+ },
2273
+ {
2274
+ "epoch": 2.990430622009569,
2275
+ "grad_norm": 13.5625,
2276
+ "learning_rate": 1.7942583732057418e-07,
2277
+ "loss": 7.5145,
2278
+ "mean_token_accuracy": 0.058688531070947646,
2279
+ "num_tokens": 3953646.0,
2280
+ "step": 2500
2281
+ },
2282
+ {
2283
+ "epoch": 3.0,
2284
+ "eval_loss": 7.677623271942139,
2285
+ "eval_mean_token_accuracy": 0.05568885817280356,
2286
+ "eval_num_tokens": 3963840.0,
2287
+ "eval_runtime": 18.6561,
2288
+ "eval_samples_per_second": 22.406,
2289
+ "eval_steps_per_second": 2.841,
2290
+ "step": 2508
2291
+ }
2292
+ ],
2293
+ "logging_steps": 10,
2294
+ "max_steps": 2508,
2295
+ "num_input_tokens_seen": 0,
2296
+ "num_train_epochs": 3,
2297
+ "save_steps": 500,
2298
+ "stateful_callbacks": {
2299
+ "TrainerControl": {
2300
+ "args": {
2301
+ "should_epoch_stop": false,
2302
+ "should_evaluate": false,
2303
+ "should_log": false,
2304
+ "should_save": true,
2305
+ "should_training_stop": true
2306
+ },
2307
+ "attributes": {}
2308
+ }
2309
+ },
2310
+ "total_flos": 1.363266938316718e+17,
2311
+ "train_batch_size": 4,
2312
+ "trial_name": null,
2313
+ "trial_params": null
2314
+ }
sft_output_simpleQA/checkpoint-2508/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:010b0827320093053654035df7d8f073bcd42c7bac7c3dea5c34279a14edc2af
3
+ size 5688